[INFO] cloning repository https://github.com/YusukeSuzuki160/minsv-parser [INFO] running `Command { std: "git" "-c" "credential.helper=" "-c" "credential.helper=/workspace/cargo-home/bin/git-credential-null" "clone" "--bare" "https://github.com/YusukeSuzuki160/minsv-parser" "/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2FYusukeSuzuki160%2Fminsv-parser", kill_on_drop: false }` [INFO] [stderr] Cloning into bare repository '/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2FYusukeSuzuki160%2Fminsv-parser'... [INFO] running `Command { std: "git" "rev-parse" "HEAD", kill_on_drop: false }` [INFO] [stdout] b982dfe70a46a032ee7e85d25428e37d7b2f6d4c [INFO] building YusukeSuzuki160/minsv-parser against master#11663cd3bfefef7d34e8f0892c250bf698049392+rustflags=-Dtail-expr-drop-order for pr-134523 [INFO] running `Command { std: "git" "clone" "/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2FYusukeSuzuki160%2Fminsv-parser" "/workspace/builds/worker-5-tc1/source", kill_on_drop: false }` [INFO] [stderr] Cloning into '/workspace/builds/worker-5-tc1/source'... [INFO] [stderr] done. [INFO] validating manifest of git repo https://github.com/YusukeSuzuki160/minsv-parser on toolchain 11663cd3bfefef7d34e8f0892c250bf698049392 [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+11663cd3bfefef7d34e8f0892c250bf698049392" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }` [INFO] started tweaking git repo https://github.com/YusukeSuzuki160/minsv-parser [INFO] finished tweaking git repo https://github.com/YusukeSuzuki160/minsv-parser [INFO] tweaked toml for git repo https://github.com/YusukeSuzuki160/minsv-parser written to /workspace/builds/worker-5-tc1/source/Cargo.toml [INFO] crate git repo https://github.com/YusukeSuzuki160/minsv-parser already has a lockfile, it will not be regenerated [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+11663cd3bfefef7d34e8f0892c250bf698049392" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }` [INFO] [stderr] Blocking waiting for file lock on package cache [INFO] [stderr] Blocking waiting for file lock on package cache [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:f3a9d4ad9d972b27faf3965f35b62e55ba32bbce8f20bc8fe909558a86702fde" "/opt/rustwide/cargo-home/bin/cargo" "+11663cd3bfefef7d34e8f0892c250bf698049392" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }` [INFO] [stdout] caf8b73702fed6aa73b59ccadf13b677ca9a4b9e75a78ebdab3cb6573c9036ac [INFO] running `Command { std: "docker" "start" "-a" "caf8b73702fed6aa73b59ccadf13b677ca9a4b9e75a78ebdab3cb6573c9036ac", kill_on_drop: false }` [INFO] running `Command { std: "docker" "inspect" "caf8b73702fed6aa73b59ccadf13b677ca9a4b9e75a78ebdab3cb6573c9036ac", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "caf8b73702fed6aa73b59ccadf13b677ca9a4b9e75a78ebdab3cb6573c9036ac", kill_on_drop: false }` [INFO] [stdout] caf8b73702fed6aa73b59ccadf13b677ca9a4b9e75a78ebdab3cb6573c9036ac [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid -Dtail-expr-drop-order" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:f3a9d4ad9d972b27faf3965f35b62e55ba32bbce8f20bc8fe909558a86702fde" "/opt/rustwide/cargo-home/bin/cargo" "+11663cd3bfefef7d34e8f0892c250bf698049392" "build" "--frozen" "--message-format=json", kill_on_drop: false }` [INFO] [stdout] 16e61ba5dd373fc6ec14e328ff4d4d5f941bf5105380b4e1d74be92865ba895e [INFO] running `Command { std: "docker" "start" "-a" "16e61ba5dd373fc6ec14e328ff4d4d5f941bf5105380b4e1d74be92865ba895e", kill_on_drop: false }` [INFO] [stderr] Compiling memchr v2.5.0 [INFO] [stderr] Compiling minimal-lexical v0.2.1 [INFO] [stderr] Compiling nom v7.1.3 [INFO] [stderr] Compiling minsv-parser v0.1.0 (/opt/rustwide/workdir) [INFO] [stdout] warning: unused imports: `is_a`, `many1`, `opt`, `pair`, `recognize`, `take_while`, and `value` [INFO] [stdout] --> src/lexer.rs:3:23 [INFO] [stdout] | [INFO] [stdout] 3 | bytes::complete::{is_a, tag, take_while, take_while1}, [INFO] [stdout] | ^^^^ ^^^^^^^^^^ [INFO] [stdout] 4 | character::complete::multispace0, [INFO] [stdout] 5 | combinator::{map, opt, recognize, value}, [INFO] [stdout] | ^^^ ^^^^^^^^^ ^^^^^ [INFO] [stdout] 6 | multi::{many0, many1}, [INFO] [stdout] | ^^^^^ [INFO] [stdout] 7 | sequence::{delimited, pair}, [INFO] [stdout] | ^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variant `NOT_EQ` should have an upper camel case name [INFO] [stdout] --> src/lexer.rs:39:5 [INFO] [stdout] | [INFO] [stdout] 39 | NOT_EQ, [INFO] [stdout] | ^^^^^^ help: convert the identifier to upper camel case: `NotEq` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(non_camel_case_types)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variant `LT_EQ` should have an upper camel case name [INFO] [stdout] --> src/lexer.rs:40:5 [INFO] [stdout] | [INFO] [stdout] 40 | LT_EQ, [INFO] [stdout] | ^^^^^ help: convert the identifier to upper camel case: `LtEq` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variant `GT_EQ` should have an upper camel case name [INFO] [stdout] --> src/lexer.rs:41:5 [INFO] [stdout] | [INFO] [stdout] 41 | GT_EQ, [INFO] [stdout] | ^^^^^ help: convert the identifier to upper camel case: `GtEq` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `identifier` is never used [INFO] [stdout] --> src/lexer.rs:61:4 [INFO] [stdout] | [INFO] [stdout] 61 | fn identifier(input: &str) -> IResult<&str, &str> { [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(dead_code)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_identifier` is never used [INFO] [stdout] --> src/lexer.rs:65:4 [INFO] [stdout] | [INFO] [stdout] 65 | fn token_identifier(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `is_number` is never used [INFO] [stdout] --> src/lexer.rs:93:4 [INFO] [stdout] | [INFO] [stdout] 93 | fn is_number(s: &str) -> bool { [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_comma` is never used [INFO] [stdout] --> src/lexer.rs:99:4 [INFO] [stdout] | [INFO] [stdout] 99 | fn token_comma(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_semicolon` is never used [INFO] [stdout] --> src/lexer.rs:103:4 [INFO] [stdout] | [INFO] [stdout] 103 | fn token_semicolon(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_lparen` is never used [INFO] [stdout] --> src/lexer.rs:107:4 [INFO] [stdout] | [INFO] [stdout] 107 | fn token_lparen(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_rparen` is never used [INFO] [stdout] --> src/lexer.rs:111:4 [INFO] [stdout] | [INFO] [stdout] 111 | fn token_rparen(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_lbracket` is never used [INFO] [stdout] --> src/lexer.rs:115:4 [INFO] [stdout] | [INFO] [stdout] 115 | fn token_lbracket(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_rbracket` is never used [INFO] [stdout] --> src/lexer.rs:119:4 [INFO] [stdout] | [INFO] [stdout] 119 | fn token_rbracket(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_plus` is never used [INFO] [stdout] --> src/lexer.rs:123:4 [INFO] [stdout] | [INFO] [stdout] 123 | fn token_plus(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_minus` is never used [INFO] [stdout] --> src/lexer.rs:127:4 [INFO] [stdout] | [INFO] [stdout] 127 | fn token_minus(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_star` is never used [INFO] [stdout] --> src/lexer.rs:131:4 [INFO] [stdout] | [INFO] [stdout] 131 | fn token_star(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_slash` is never used [INFO] [stdout] --> src/lexer.rs:135:4 [INFO] [stdout] | [INFO] [stdout] 135 | fn token_slash(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_percent` is never used [INFO] [stdout] --> src/lexer.rs:139:4 [INFO] [stdout] | [INFO] [stdout] 139 | fn token_percent(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_ampersand` is never used [INFO] [stdout] --> src/lexer.rs:143:4 [INFO] [stdout] | [INFO] [stdout] 143 | fn token_ampersand(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_pipe` is never used [INFO] [stdout] --> src/lexer.rs:147:4 [INFO] [stdout] | [INFO] [stdout] 147 | fn token_pipe(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_caret` is never used [INFO] [stdout] --> src/lexer.rs:151:4 [INFO] [stdout] | [INFO] [stdout] 151 | fn token_caret(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_lt` is never used [INFO] [stdout] --> src/lexer.rs:155:4 [INFO] [stdout] | [INFO] [stdout] 155 | fn token_lt(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_gt` is never used [INFO] [stdout] --> src/lexer.rs:159:4 [INFO] [stdout] | [INFO] [stdout] 159 | fn token_gt(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_eq` is never used [INFO] [stdout] --> src/lexer.rs:163:4 [INFO] [stdout] | [INFO] [stdout] 163 | fn token_eq(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_not_eq` is never used [INFO] [stdout] --> src/lexer.rs:167:4 [INFO] [stdout] | [INFO] [stdout] 167 | fn token_not_eq(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_lt_eq` is never used [INFO] [stdout] --> src/lexer.rs:171:4 [INFO] [stdout] | [INFO] [stdout] 171 | fn token_lt_eq(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_gt_eq` is never used [INFO] [stdout] --> src/lexer.rs:175:4 [INFO] [stdout] | [INFO] [stdout] 175 | fn token_gt_eq(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_and` is never used [INFO] [stdout] --> src/lexer.rs:179:4 [INFO] [stdout] | [INFO] [stdout] 179 | fn token_and(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_or` is never used [INFO] [stdout] --> src/lexer.rs:183:4 [INFO] [stdout] | [INFO] [stdout] 183 | fn token_or(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_lshift` is never used [INFO] [stdout] --> src/lexer.rs:187:4 [INFO] [stdout] | [INFO] [stdout] 187 | fn token_lshift(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_rshift` is never used [INFO] [stdout] --> src/lexer.rs:191:4 [INFO] [stdout] | [INFO] [stdout] 191 | fn token_rshift(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_nonblcok_assign` is never used [INFO] [stdout] --> src/lexer.rs:195:4 [INFO] [stdout] | [INFO] [stdout] 195 | fn token_nonblcok_assign(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_block_assign` is never used [INFO] [stdout] --> src/lexer.rs:199:4 [INFO] [stdout] | [INFO] [stdout] 199 | fn token_block_assign(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_colon` is never used [INFO] [stdout] --> src/lexer.rs:203:4 [INFO] [stdout] | [INFO] [stdout] 203 | fn token_colon(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_at` is never used [INFO] [stdout] --> src/lexer.rs:207:4 [INFO] [stdout] | [INFO] [stdout] 207 | fn token_at(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_begin` is never used [INFO] [stdout] --> src/lexer.rs:211:4 [INFO] [stdout] | [INFO] [stdout] 211 | fn token_begin(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_end` is never used [INFO] [stdout] --> src/lexer.rs:215:4 [INFO] [stdout] | [INFO] [stdout] 215 | fn token_end(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_if` is never used [INFO] [stdout] --> src/lexer.rs:219:4 [INFO] [stdout] | [INFO] [stdout] 219 | fn token_if(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_else` is never used [INFO] [stdout] --> src/lexer.rs:223:4 [INFO] [stdout] | [INFO] [stdout] 223 | fn token_else(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_always` is never used [INFO] [stdout] --> src/lexer.rs:227:4 [INFO] [stdout] | [INFO] [stdout] 227 | fn token_always(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_posedge` is never used [INFO] [stdout] --> src/lexer.rs:231:4 [INFO] [stdout] | [INFO] [stdout] 231 | fn token_posedge(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_negedge` is never used [INFO] [stdout] --> src/lexer.rs:235:4 [INFO] [stdout] | [INFO] [stdout] 235 | fn token_negedge(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `whitespace` is never used [INFO] [stdout] --> src/lexer.rs:239:4 [INFO] [stdout] | [INFO] [stdout] 239 | fn whitespace<'a, F, O, E: ParseError<&'a str>>(inner: F) -> impl FnMut(&'a str) -> IResult<&'a str, O, E> [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token` is never used [INFO] [stdout] --> src/lexer.rs:250:4 [INFO] [stdout] | [INFO] [stdout] 250 | fn token(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `tokens` is never used [INFO] [stdout] --> src/lexer.rs:291:4 [INFO] [stdout] | [INFO] [stdout] 291 | fn tokens(input: &str) -> IResult<&str, Vec> { [INFO] [stdout] | ^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: struct `Node` is never constructed [INFO] [stdout] --> src/parser.rs:4:8 [INFO] [stdout] | [INFO] [stdout] 4 | struct Node<'a> { [INFO] [stdout] | ^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: struct `Ast` is never constructed [INFO] [stdout] --> src/parser.rs:10:8 [INFO] [stdout] | [INFO] [stdout] 10 | struct Ast<'a> { [INFO] [stdout] | ^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] Finished `dev` profile [unoptimized + debuginfo] target(s) in 2.59s [INFO] running `Command { std: "docker" "inspect" "16e61ba5dd373fc6ec14e328ff4d4d5f941bf5105380b4e1d74be92865ba895e", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "16e61ba5dd373fc6ec14e328ff4d4d5f941bf5105380b4e1d74be92865ba895e", kill_on_drop: false }` [INFO] [stdout] 16e61ba5dd373fc6ec14e328ff4d4d5f941bf5105380b4e1d74be92865ba895e [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid -Dtail-expr-drop-order" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:f3a9d4ad9d972b27faf3965f35b62e55ba32bbce8f20bc8fe909558a86702fde" "/opt/rustwide/cargo-home/bin/cargo" "+11663cd3bfefef7d34e8f0892c250bf698049392" "test" "--frozen" "--no-run" "--message-format=json", kill_on_drop: false }` [INFO] [stdout] fb47f3a5915f6e31f699fdd868479a2a37b440dd96d51e77d7cd1a962ea2a22c [INFO] running `Command { std: "docker" "start" "-a" "fb47f3a5915f6e31f699fdd868479a2a37b440dd96d51e77d7cd1a962ea2a22c", kill_on_drop: false }` [INFO] [stdout] warning: unused imports: `is_a`, `many1`, `opt`, `pair`, `recognize`, `take_while`, and `value` [INFO] [stdout] --> src/lexer.rs:3:23 [INFO] [stdout] | [INFO] [stdout] 3 | bytes::complete::{is_a, tag, take_while, take_while1}, [INFO] [stdout] | ^^^^ ^^^^^^^^^^ [INFO] [stdout] 4 | character::complete::multispace0, [INFO] [stdout] 5 | combinator::{map, opt, recognize, value}, [INFO] [stdout] | ^^^ ^^^^^^^^^ ^^^^^ [INFO] [stdout] 6 | multi::{many0, many1}, [INFO] [stdout] | ^^^^^ [INFO] [stdout] 7 | sequence::{delimited, pair}, [INFO] [stdout] | ^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variant `NOT_EQ` should have an upper camel case name [INFO] [stdout] --> src/lexer.rs:39:5 [INFO] [stdout] | [INFO] [stdout] 39 | NOT_EQ, [INFO] [stdout] | ^^^^^^ help: convert the identifier to upper camel case: `NotEq` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(non_camel_case_types)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variant `LT_EQ` should have an upper camel case name [INFO] [stdout] --> src/lexer.rs:40:5 [INFO] [stdout] | [INFO] [stdout] 40 | LT_EQ, [INFO] [stdout] | ^^^^^ help: convert the identifier to upper camel case: `LtEq` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variant `GT_EQ` should have an upper camel case name [INFO] [stdout] --> src/lexer.rs:41:5 [INFO] [stdout] | [INFO] [stdout] 41 | GT_EQ, [INFO] [stdout] | ^^^^^ help: convert the identifier to upper camel case: `GtEq` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `identifier` is never used [INFO] [stdout] --> src/lexer.rs:61:4 [INFO] [stdout] | [INFO] [stdout] 61 | fn identifier(input: &str) -> IResult<&str, &str> { [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(dead_code)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_identifier` is never used [INFO] [stdout] --> src/lexer.rs:65:4 [INFO] [stdout] | [INFO] [stdout] 65 | fn token_identifier(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `is_number` is never used [INFO] [stdout] --> src/lexer.rs:93:4 [INFO] [stdout] | [INFO] [stdout] 93 | fn is_number(s: &str) -> bool { [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_comma` is never used [INFO] [stdout] --> src/lexer.rs:99:4 [INFO] [stdout] | [INFO] [stdout] 99 | fn token_comma(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_semicolon` is never used [INFO] [stdout] --> src/lexer.rs:103:4 [INFO] [stdout] | [INFO] [stdout] 103 | fn token_semicolon(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_lparen` is never used [INFO] [stdout] --> src/lexer.rs:107:4 [INFO] [stdout] | [INFO] [stdout] 107 | fn token_lparen(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_rparen` is never used [INFO] [stdout] --> src/lexer.rs:111:4 [INFO] [stdout] | [INFO] [stdout] 111 | fn token_rparen(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_lbracket` is never used [INFO] [stdout] --> src/lexer.rs:115:4 [INFO] [stdout] | [INFO] [stdout] 115 | fn token_lbracket(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_rbracket` is never used [INFO] [stdout] --> src/lexer.rs:119:4 [INFO] [stdout] | [INFO] [stdout] 119 | fn token_rbracket(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_plus` is never used [INFO] [stdout] --> src/lexer.rs:123:4 [INFO] [stdout] | [INFO] [stdout] 123 | fn token_plus(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_minus` is never used [INFO] [stdout] --> src/lexer.rs:127:4 [INFO] [stdout] | [INFO] [stdout] 127 | fn token_minus(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_star` is never used [INFO] [stdout] --> src/lexer.rs:131:4 [INFO] [stdout] | [INFO] [stdout] 131 | fn token_star(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_slash` is never used [INFO] [stdout] --> src/lexer.rs:135:4 [INFO] [stdout] | [INFO] [stdout] 135 | fn token_slash(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_percent` is never used [INFO] [stdout] --> src/lexer.rs:139:4 [INFO] [stdout] | [INFO] [stdout] 139 | fn token_percent(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_ampersand` is never used [INFO] [stdout] --> src/lexer.rs:143:4 [INFO] [stdout] | [INFO] [stdout] 143 | fn token_ampersand(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_pipe` is never used [INFO] [stdout] --> src/lexer.rs:147:4 [INFO] [stdout] | [INFO] [stdout] 147 | fn token_pipe(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_caret` is never used [INFO] [stdout] --> src/lexer.rs:151:4 [INFO] [stdout] | [INFO] [stdout] 151 | fn token_caret(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_lt` is never used [INFO] [stdout] --> src/lexer.rs:155:4 [INFO] [stdout] | [INFO] [stdout] 155 | fn token_lt(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_gt` is never used [INFO] [stdout] --> src/lexer.rs:159:4 [INFO] [stdout] | [INFO] [stdout] 159 | fn token_gt(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_eq` is never used [INFO] [stdout] --> src/lexer.rs:163:4 [INFO] [stdout] | [INFO] [stdout] 163 | fn token_eq(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_not_eq` is never used [INFO] [stdout] --> src/lexer.rs:167:4 [INFO] [stdout] | [INFO] [stdout] 167 | fn token_not_eq(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_lt_eq` is never used [INFO] [stdout] --> src/lexer.rs:171:4 [INFO] [stdout] | [INFO] [stdout] 171 | fn token_lt_eq(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_gt_eq` is never used [INFO] [stdout] --> src/lexer.rs:175:4 [INFO] [stdout] | [INFO] [stdout] 175 | fn token_gt_eq(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_and` is never used [INFO] [stdout] --> src/lexer.rs:179:4 [INFO] [stdout] | [INFO] [stdout] 179 | fn token_and(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_or` is never used [INFO] [stdout] --> src/lexer.rs:183:4 [INFO] [stdout] | [INFO] [stdout] 183 | fn token_or(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_lshift` is never used [INFO] [stdout] --> src/lexer.rs:187:4 [INFO] [stdout] | [INFO] [stdout] 187 | fn token_lshift(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_rshift` is never used [INFO] [stdout] --> src/lexer.rs:191:4 [INFO] [stdout] | [INFO] [stdout] 191 | fn token_rshift(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_nonblcok_assign` is never used [INFO] [stdout] --> src/lexer.rs:195:4 [INFO] [stdout] | [INFO] [stdout] 195 | fn token_nonblcok_assign(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_block_assign` is never used [INFO] [stdout] --> src/lexer.rs:199:4 [INFO] [stdout] | [INFO] [stdout] 199 | fn token_block_assign(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_colon` is never used [INFO] [stdout] --> src/lexer.rs:203:4 [INFO] [stdout] | [INFO] [stdout] 203 | fn token_colon(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_at` is never used [INFO] [stdout] --> src/lexer.rs:207:4 [INFO] [stdout] | [INFO] [stdout] 207 | fn token_at(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_begin` is never used [INFO] [stdout] --> src/lexer.rs:211:4 [INFO] [stdout] | [INFO] [stdout] 211 | fn token_begin(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_end` is never used [INFO] [stdout] --> src/lexer.rs:215:4 [INFO] [stdout] | [INFO] [stdout] 215 | fn token_end(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_if` is never used [INFO] [stdout] --> src/lexer.rs:219:4 [INFO] [stdout] | [INFO] [stdout] 219 | fn token_if(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] Compiling minsv-parser v0.1.0 (/opt/rustwide/workdir) [INFO] [stdout] warning: function `token_else` is never used [INFO] [stdout] --> src/lexer.rs:223:4 [INFO] [stdout] | [INFO] [stdout] 223 | fn token_else(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_always` is never used [INFO] [stdout] --> src/lexer.rs:227:4 [INFO] [stdout] | [INFO] [stdout] 227 | fn token_always(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_posedge` is never used [INFO] [stdout] --> src/lexer.rs:231:4 [INFO] [stdout] | [INFO] [stdout] 231 | fn token_posedge(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token_negedge` is never used [INFO] [stdout] --> src/lexer.rs:235:4 [INFO] [stdout] | [INFO] [stdout] 235 | fn token_negedge(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `whitespace` is never used [INFO] [stdout] --> src/lexer.rs:239:4 [INFO] [stdout] | [INFO] [stdout] 239 | fn whitespace<'a, F, O, E: ParseError<&'a str>>(inner: F) -> impl FnMut(&'a str) -> IResult<&'a str, O, E> [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `token` is never used [INFO] [stdout] --> src/lexer.rs:250:4 [INFO] [stdout] | [INFO] [stdout] 250 | fn token(input: &str) -> IResult<&str, Token> { [INFO] [stdout] | ^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `tokens` is never used [INFO] [stdout] --> src/lexer.rs:291:4 [INFO] [stdout] | [INFO] [stdout] 291 | fn tokens(input: &str) -> IResult<&str, Vec> { [INFO] [stdout] | ^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: struct `Node` is never constructed [INFO] [stdout] --> src/parser.rs:4:8 [INFO] [stdout] | [INFO] [stdout] 4 | struct Node<'a> { [INFO] [stdout] | ^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: struct `Ast` is never constructed [INFO] [stdout] --> src/parser.rs:10:8 [INFO] [stdout] | [INFO] [stdout] 10 | struct Ast<'a> { [INFO] [stdout] | ^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused imports: `is_a`, `many1`, `opt`, `pair`, `recognize`, `take_while`, and `value` [INFO] [stdout] --> src/lexer.rs:3:23 [INFO] [stdout] | [INFO] [stdout] 3 | bytes::complete::{is_a, tag, take_while, take_while1}, [INFO] [stdout] | ^^^^ ^^^^^^^^^^ [INFO] [stdout] 4 | character::complete::multispace0, [INFO] [stdout] 5 | combinator::{map, opt, recognize, value}, [INFO] [stdout] | ^^^ ^^^^^^^^^ ^^^^^ [INFO] [stdout] 6 | multi::{many0, many1}, [INFO] [stdout] | ^^^^^ [INFO] [stdout] 7 | sequence::{delimited, pair}, [INFO] [stdout] | ^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variant `NOT_EQ` should have an upper camel case name [INFO] [stdout] --> src/lexer.rs:39:5 [INFO] [stdout] | [INFO] [stdout] 39 | NOT_EQ, [INFO] [stdout] | ^^^^^^ help: convert the identifier to upper camel case: `NotEq` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(non_camel_case_types)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variant `LT_EQ` should have an upper camel case name [INFO] [stdout] --> src/lexer.rs:40:5 [INFO] [stdout] | [INFO] [stdout] 40 | LT_EQ, [INFO] [stdout] | ^^^^^ help: convert the identifier to upper camel case: `LtEq` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variant `GT_EQ` should have an upper camel case name [INFO] [stdout] --> src/lexer.rs:41:5 [INFO] [stdout] | [INFO] [stdout] 41 | GT_EQ, [INFO] [stdout] | ^^^^^ help: convert the identifier to upper camel case: `GtEq` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: struct `Node` is never constructed [INFO] [stdout] --> src/parser.rs:4:8 [INFO] [stdout] | [INFO] [stdout] 4 | struct Node<'a> { [INFO] [stdout] | ^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(dead_code)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: struct `Ast` is never constructed [INFO] [stdout] --> src/parser.rs:10:8 [INFO] [stdout] | [INFO] [stdout] 10 | struct Ast<'a> { [INFO] [stdout] | ^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] Finished `test` profile [unoptimized + debuginfo] target(s) in 0.39s [INFO] running `Command { std: "docker" "inspect" "fb47f3a5915f6e31f699fdd868479a2a37b440dd96d51e77d7cd1a962ea2a22c", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "fb47f3a5915f6e31f699fdd868479a2a37b440dd96d51e77d7cd1a962ea2a22c", kill_on_drop: false }` [INFO] [stdout] fb47f3a5915f6e31f699fdd868479a2a37b440dd96d51e77d7cd1a962ea2a22c