[INFO] cloning repository https://github.com/Eucladia/toy_language_compiler
[INFO] running `Command { std: "git" "-c" "credential.helper=" "-c" "credential.helper=/workspace/cargo-home/bin/git-credential-null" "clone" "--bare" "https://github.com/Eucladia/toy_language_compiler" "/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2FEucladia%2Ftoy_language_compiler", kill_on_drop: false }`
[INFO] [stderr] Cloning into bare repository '/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2FEucladia%2Ftoy_language_compiler'...
[INFO] running `Command { std: "git" "rev-parse" "HEAD", kill_on_drop: false }`
[INFO] [stdout] 18e05474b19cb9eadcb4d78b67539a7059a278a4
[INFO] checking Eucladia/toy_language_compiler against master#8f21a5c92ea55c348c275a1bc4fedbdf181e0d64 for pr-143011
[INFO] running `Command { std: "git" "clone" "/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2FEucladia%2Ftoy_language_compiler" "/workspace/builds/worker-7-tc1/source", kill_on_drop: false }`
[INFO] [stderr] Cloning into '/workspace/builds/worker-7-tc1/source'...
[INFO] [stderr] done.
[INFO] started tweaking git repo https://github.com/Eucladia/toy_language_compiler
[INFO] finished tweaking git repo https://github.com/Eucladia/toy_language_compiler
[INFO] tweaked toml for git repo https://github.com/Eucladia/toy_language_compiler written to /workspace/builds/worker-7-tc1/source/Cargo.toml
[INFO] validating manifest of git repo https://github.com/Eucladia/toy_language_compiler on toolchain 8f21a5c92ea55c348c275a1bc4fedbdf181e0d64
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+8f21a5c92ea55c348c275a1bc4fedbdf181e0d64" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }`
[INFO] crate git repo https://github.com/Eucladia/toy_language_compiler already has a lockfile, it will not be regenerated
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+8f21a5c92ea55c348c275a1bc4fedbdf181e0d64" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-7-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-7-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:90999bfc7ae267e83380e433d8e61a7c072ca6729e92edbae886d3423b3a6f4c" "/opt/rustwide/cargo-home/bin/cargo" "+8f21a5c92ea55c348c275a1bc4fedbdf181e0d64" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }`
[INFO] [stdout] 27911057b4c1b2a948132bc7162ac72674aae8c67225da0654e68385cd0a068a
[INFO] running `Command { std: "docker" "start" "-a" "27911057b4c1b2a948132bc7162ac72674aae8c67225da0654e68385cd0a068a", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "inspect" "27911057b4c1b2a948132bc7162ac72674aae8c67225da0654e68385cd0a068a", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "27911057b4c1b2a948132bc7162ac72674aae8c67225da0654e68385cd0a068a", kill_on_drop: false }`
[INFO] [stdout] 27911057b4c1b2a948132bc7162ac72674aae8c67225da0654e68385cd0a068a
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-7-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-7-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:90999bfc7ae267e83380e433d8e61a7c072ca6729e92edbae886d3423b3a6f4c" "/opt/rustwide/cargo-home/bin/cargo" "+8f21a5c92ea55c348c275a1bc4fedbdf181e0d64" "check" "--frozen" "--all" "--all-targets" "--message-format=json", kill_on_drop: false }`
[INFO] [stdout] 930d2ef8659e90e0c088f2a3d946613d0875656fbaa066e875d30dabfb84c2ff
[INFO] running `Command { std: "docker" "start" "-a" "930d2ef8659e90e0c088f2a3d946613d0875656fbaa066e875d30dabfb84c2ff", kill_on_drop: false }`
[INFO] [stderr]     Checking toy_language v0.1.0 (/opt/rustwide/workdir)
[INFO] [stdout] error: couldn't read `src/../sample_input/1.txt`: No such file or directory (os error 2)
[INFO] [stdout]    --> src/lexer.rs:224:30
[INFO] [stdout]     |
[INFO] [stdout] 224 |       let tokens = get_tokens!(include_str!(concat!("../sample_input/", $src, ".txt")));
[INFO] [stdout]     |                                ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout] 271 | /     are_tokens_equal!(
[INFO] [stdout] 272 | |       "1",
[INFO] [stdout] 273 | |        Identifier, Equal, Literal, Semicolon
[INFO] [stdout] 274 | |     );
[INFO] [stdout]     | |_____- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this error originates in the macro `include_str` which comes from the expansion of the macro `are_tokens_equal` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: couldn't read `src/../sample_input/2.txt`: No such file or directory (os error 2)
[INFO] [stdout]    --> src/lexer.rs:224:30
[INFO] [stdout]     |
[INFO] [stdout] 224 |       let tokens = get_tokens!(include_str!(concat!("../sample_input/", $src, ".txt")));
[INFO] [stdout]     |                                ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout] 280 | /     are_tokens_equal!(
[INFO] [stdout] 281 | |       "2",
[INFO] [stdout] 282 | |        Identifier, Equal, Literal, Semicolon
[INFO] [stdout] 283 | |     );
[INFO] [stdout]     | |_____- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this error originates in the macro `include_str` which comes from the expansion of the macro `are_tokens_equal` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: couldn't read `src/../sample_input/3.txt`: No such file or directory (os error 2)
[INFO] [stdout]    --> src/lexer.rs:224:30
[INFO] [stdout]     |
[INFO] [stdout] 224 |       let tokens = get_tokens!(include_str!(concat!("../sample_input/", $src, ".txt")));
[INFO] [stdout]     |                                ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout] 289 | /     are_tokens_equal!(
[INFO] [stdout] 290 | |       "3",
[INFO] [stdout] 291 | |        Identifier, Equal, Literal,
[INFO] [stdout] 292 | |        Identifier, Equal, Identifier, Semicolon,
[INFO] [stdout] 293 | |        Identifier, Equal, Minus, Minus, Minus, LeftParen, Identifier, Plus, Identifier, RightParen, Semicolon
[INFO] [stdout] 294 | |     );
[INFO] [stdout]     | |_____- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this error originates in the macro `include_str` which comes from the expansion of the macro `are_tokens_equal` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: couldn't read `src/../sample_input/4.txt`: No such file or directory (os error 2)
[INFO] [stdout]    --> src/lexer.rs:224:30
[INFO] [stdout]     |
[INFO] [stdout] 224 |       let tokens = get_tokens!(include_str!(concat!("../sample_input/", $src, ".txt")));
[INFO] [stdout]     |                                ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout] 300 | /     are_tokens_equal!(
[INFO] [stdout] 301 | |       "4",
[INFO] [stdout] 302 | |       Identifier, Equal, Literal, Semicolon,
[INFO] [stdout] 303 | |       Identifier, Equal, Literal, Semicolon,
[INFO] [stdout] 304 | |       Identifier, Equal, Minus, Minus, Minus, LeftParen, Identifier, Plus, Identifi...
[INFO] [stdout] 305 | |     );
[INFO] [stdout]     | |_____- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this error originates in the macro `include_str` which comes from the expansion of the macro `are_tokens_equal` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stderr] error: could not compile `toy_language` (bin "toy_language" test) due to 4 previous errors
[INFO] running `Command { std: "docker" "inspect" "930d2ef8659e90e0c088f2a3d946613d0875656fbaa066e875d30dabfb84c2ff", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "930d2ef8659e90e0c088f2a3d946613d0875656fbaa066e875d30dabfb84c2ff", kill_on_drop: false }`
[INFO] [stdout] 930d2ef8659e90e0c088f2a3d946613d0875656fbaa066e875d30dabfb84c2ff
