[INFO] crate luthor 0.2.0 is already in cache [INFO] extracting crate luthor 0.2.0 into work/ex/beta-1.37-6/sources/1.36.0/reg/luthor/0.2.0 [INFO] extracting crate luthor 0.2.0 into work/ex/beta-1.37-6/sources/beta-2019-07-23/reg/luthor/0.2.0 [INFO] validating manifest of luthor-0.2.0 on toolchain 1.36.0 [INFO] running `"/mnt/big/crater/work/local/cargo-home/bin/cargo" "+1.36.0" "read-manifest" "--manifest-path" "Cargo.toml"` [INFO] validating manifest of luthor-0.2.0 on toolchain beta-2019-07-23 [INFO] running `"/mnt/big/crater/work/local/cargo-home/bin/cargo" "+beta-2019-07-23" "read-manifest" "--manifest-path" "Cargo.toml"` [INFO] started frobbing luthor-0.2.0 [INFO] finished frobbing luthor-0.2.0 [INFO] frobbed toml for luthor-0.2.0 written to work/ex/beta-1.37-6/sources/1.36.0/reg/luthor/0.2.0/Cargo.toml [INFO] started frobbing luthor-0.2.0 [INFO] finished frobbing luthor-0.2.0 [INFO] frobbed toml for luthor-0.2.0 written to work/ex/beta-1.37-6/sources/beta-2019-07-23/reg/luthor/0.2.0/Cargo.toml [INFO] running `"/mnt/big/crater/work/local/cargo-home/bin/cargo" "+1.36.0" "generate-lockfile" "--manifest-path" "Cargo.toml" "-Zno-index-update"` [INFO] running `"/mnt/big/crater/work/local/cargo-home/bin/cargo" "+beta-2019-07-23" "generate-lockfile" "--manifest-path" "Cargo.toml" "-Zno-index-update"` [INFO] running `"/mnt/big/crater/work/local/cargo-home/bin/cargo" "+1.36.0" "fetch" "--locked" "--manifest-path" "Cargo.toml"` [INFO] running `"/mnt/big/crater/work/local/cargo-home/bin/cargo" "+beta-2019-07-23" "fetch" "--locked" "--manifest-path" "Cargo.toml"` [INFO] testing luthor-0.2.0 against beta-2019-07-23 for beta-1.37-6 [INFO] running `"docker" "create" "-v" "/mnt/big/crater/work/local/target-dirs/beta-1.37-6/worker-5/beta-2019-07-23:/opt/crater/target:rw,Z" "-v" "/mnt/big/crater/work/ex/beta-1.37-6/sources/beta-2019-07-23/reg/luthor/0.2.0:/opt/crater/workdir:ro,Z" "-v" "/mnt/big/crater/work/local/cargo-home:/opt/crater/cargo-home:ro,Z" "-v" "/mnt/big/crater/work/local/rustup-home:/opt/crater/rustup-home:ro,Z" "-e" "USER_ID=1000" "-e" "SOURCE_DIR=/opt/crater/workdir" "-e" "MAP_USER_ID=1000" "-e" "CARGO_TARGET_DIR=/opt/crater/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=warn" "-e" "CARGO_HOME=/opt/crater/cargo-home" "-e" "RUSTUP_HOME=/opt/crater/rustup-home" "-w" "/opt/crater/workdir" "-m" "1536M" "--network" "none" "rustops/crates-build-env" "/opt/crater/cargo-home/bin/cargo" "+beta-2019-07-23" "build" "--frozen"` [INFO] [stdout] e8aed62f8127b2416f8b63f32faeaa54500dc1d1fcdfa8c2576113ca32401687 [INFO] running `"docker" "start" "-a" "e8aed62f8127b2416f8b63f32faeaa54500dc1d1fcdfa8c2576113ca32401687"` [INFO] [stderr] Compiling luthor v0.2.0 (/opt/crater/workdir) [INFO] [stderr] Finished dev [unoptimized + debuginfo] target(s) in 0.95s [INFO] running `"docker" "inspect" "e8aed62f8127b2416f8b63f32faeaa54500dc1d1fcdfa8c2576113ca32401687"` [INFO] running `"docker" "rm" "-f" "e8aed62f8127b2416f8b63f32faeaa54500dc1d1fcdfa8c2576113ca32401687"` [INFO] [stdout] e8aed62f8127b2416f8b63f32faeaa54500dc1d1fcdfa8c2576113ca32401687 [INFO] running `"docker" "create" "-v" "/mnt/big/crater/work/local/target-dirs/beta-1.37-6/worker-5/beta-2019-07-23:/opt/crater/target:rw,Z" "-v" "/mnt/big/crater/work/ex/beta-1.37-6/sources/beta-2019-07-23/reg/luthor/0.2.0:/opt/crater/workdir:ro,Z" "-v" "/mnt/big/crater/work/local/cargo-home:/opt/crater/cargo-home:ro,Z" "-v" "/mnt/big/crater/work/local/rustup-home:/opt/crater/rustup-home:ro,Z" "-e" "USER_ID=1000" "-e" "SOURCE_DIR=/opt/crater/workdir" "-e" "MAP_USER_ID=1000" "-e" "CARGO_TARGET_DIR=/opt/crater/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=warn" "-e" "CARGO_HOME=/opt/crater/cargo-home" "-e" "RUSTUP_HOME=/opt/crater/rustup-home" "-w" "/opt/crater/workdir" "-m" "1536M" "--network" "none" "rustops/crates-build-env" "/opt/crater/cargo-home/bin/cargo" "+beta-2019-07-23" "test" "--frozen" "--no-run"` [INFO] [stdout] 932b58e158270804fbfeb8369af5d7da35b8e89fb3c39d298ddbee63ed95d2a5 [INFO] running `"docker" "start" "-a" "932b58e158270804fbfeb8369af5d7da35b8e89fb3c39d298ddbee63ed95d2a5"` [INFO] [stderr] Compiling luthor v0.2.0 (/opt/crater/workdir) [INFO] [stderr] Finished dev [unoptimized + debuginfo] target(s) in 1.64s [INFO] running `"docker" "inspect" "932b58e158270804fbfeb8369af5d7da35b8e89fb3c39d298ddbee63ed95d2a5"` [INFO] running `"docker" "rm" "-f" "932b58e158270804fbfeb8369af5d7da35b8e89fb3c39d298ddbee63ed95d2a5"` [INFO] [stdout] 932b58e158270804fbfeb8369af5d7da35b8e89fb3c39d298ddbee63ed95d2a5 [INFO] running `"docker" "create" "-v" "/mnt/big/crater/work/local/target-dirs/beta-1.37-6/worker-5/beta-2019-07-23:/opt/crater/target:rw,Z" "-v" "/mnt/big/crater/work/ex/beta-1.37-6/sources/beta-2019-07-23/reg/luthor/0.2.0:/opt/crater/workdir:ro,Z" "-v" "/mnt/big/crater/work/local/cargo-home:/opt/crater/cargo-home:ro,Z" "-v" "/mnt/big/crater/work/local/rustup-home:/opt/crater/rustup-home:ro,Z" "-e" "USER_ID=1000" "-e" "SOURCE_DIR=/opt/crater/workdir" "-e" "MAP_USER_ID=1000" "-e" "CARGO_TARGET_DIR=/opt/crater/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=warn" "-e" "CARGO_HOME=/opt/crater/cargo-home" "-e" "RUSTUP_HOME=/opt/crater/rustup-home" "-w" "/opt/crater/workdir" "-m" "1536M" "--network" "none" "rustops/crates-build-env" "/opt/crater/cargo-home/bin/cargo" "+beta-2019-07-23" "test" "--frozen"` [INFO] [stdout] cefc581706c08dc6f9aab3cffcc91aa5d1bec3d44010e18c879b1e64ba708fa4 [INFO] running `"docker" "start" "-a" "cefc581706c08dc6f9aab3cffcc91aa5d1bec3d44010e18c879b1e64ba708fa4"` [INFO] [stderr] Finished dev [unoptimized + debuginfo] target(s) in 0.03s [INFO] [stdout] [INFO] [stderr] Running /opt/crater/target/debug/deps/luthor-5a9d1191c2c96c41 [INFO] [stdout] running 22 tests [INFO] [stdout] test lexers::rust::tests::it_works ... ok [INFO] [stdout] test lexers::ruby::tests::it_works ... ok [INFO] [stdout] test lexers::ruby::tests::it_identifies_integers_and_operators ... ok [INFO] [stdout] test lexers::json::tests::it_can_handle_utf8_data ... ok [INFO] [stdout] test lexers::default::tests::it_works ... ok [INFO] [stdout] test lexers::json::tests::it_can_handle_open_strings ... ok [INFO] [stdout] test lexers::json::tests::it_works ... ok [INFO] [stdout] test lexers::xml::tests::it_can_handle_open_strings ... ok [INFO] [stdout] test tokenizer::tests::consume_whitespace_handles_preexisting_noncategorized_chars ... ok [INFO] [stdout] test lexers::json::tests::it_can_handle_garbage ... ok [INFO] [stdout] test tokenizer::tests::current_char_returns_none_if_at_the_end ... ok [INFO] [stdout] test lexers::xml::tests::it_works ... ok [INFO] [stdout] test lexers::xml::tests::it_can_handle_utf8_data ... ok [INFO] [stdout] test tokenizer::tests::tokenize_creates_the_correct_token ... ok [INFO] [stdout] test tokenizer::tests::tokens_joins_advanced_data_with_unprocessed_data_as_text_token ... ok [INFO] [stdout] test tokenizer::tests::tokenize_does_nothing_if_range_is_empty ... ok [INFO] [stdout] test tokenizer::tests::current_char_returns_the_char_at_head ... ok [INFO] [stdout] test tokenizer::tests::tokenize_next_tokenizes_previous_data_as_text ... ok [INFO] [stdout] test tokenizer::tests::tokenize_next_tokenizes_next_x_chars ... ok [INFO] [stdout] test tokenizer::tests::tokenize_next_takes_at_most_what_is_left ... ok [INFO] [stdout] test tokenizer::tests::tokens_returns_unprocessed_data_as_text_token ... ok [INFO] [stdout] test lexers::xml::tests::it_can_handle_garbage ... ok [INFO] [stderr] Doc-tests luthor [INFO] [stdout] [INFO] [stdout] test result: ok. 22 passed; 0 failed; 0 ignored; 0 measured; 0 filtered out [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] running 10 tests [INFO] [stdout] test src/tokenizer.rs - tokenizer::Tokenizer::new (line 29) ... ok [INFO] [stdout] test src/tokenizer.rs - tokenizer::Tokenizer::starts_with_lexeme (line 191) ... ok [INFO] [stdout] test src/tokenizer.rs - tokenizer::Tokenizer::has_prefix (line 163) ... ok [INFO] [stdout] test src/tokenizer.rs - tokenizer::Tokenizer::current_char (line 115) ... ok [INFO] [stdout] test src/tokenizer.rs - tokenizer::Tokenizer::advance (line 90) ... ok [INFO] [stdout] test src/tokenizer.rs - tokenizer::Tokenizer::next_non_whitespace_char (line 139) ... ok [INFO] [stdout] test src/tokenizer.rs - tokenizer::Tokenizer::consume_whitespace (line 307) ... ok [INFO] [stdout] test src/tokenizer.rs - tokenizer::Tokenizer::tokenize (line 238) ... ok [INFO] [stdout] test src/tokenizer.rs - tokenizer::Tokenizer::tokenize_next (line 270) ... ok [INFO] [stdout] test src/tokenizer.rs - tokenizer::Tokenizer::tokens (line 48) ... ok [INFO] [stdout] [INFO] [stdout] test result: ok. 10 passed; 0 failed; 0 ignored; 0 measured; 0 filtered out [INFO] [stdout] [INFO] running `"docker" "inspect" "cefc581706c08dc6f9aab3cffcc91aa5d1bec3d44010e18c879b1e64ba708fa4"` [INFO] running `"docker" "rm" "-f" "cefc581706c08dc6f9aab3cffcc91aa5d1bec3d44010e18c879b1e64ba708fa4"` [INFO] [stdout] cefc581706c08dc6f9aab3cffcc91aa5d1bec3d44010e18c879b1e64ba708fa4