Nov 18 06:52:56.760 INFO testing luthor-0.2.0 against try#2e026d6e5d7621634147ad4c8074f1d964dbb1be for pr-55982 Nov 18 06:52:56.760 INFO running `"docker" "create" "-v" "/mnt/big/crater/work/local/target-dirs/pr-55982/worker-5/try#2e026d6e5d7621634147ad4c8074f1d964dbb1be:/target:rw,Z" "-v" "/mnt/big/crater/work/local/test-source/worker-5/pr-55982/try#2e026d6e5d7621634147ad4c8074f1d964dbb1be:/source:ro,Z" "-v" "/mnt/big/crater/work/local/cargo-home:/cargo-home:ro,Z" "-v" "/mnt/big/crater/work/local/rustup-home:/rustup-home:ro,Z" "-e" "USER_ID=1000" "-e" "SOURCE_DIR=/source" "-e" "USER_ID=1000" "-e" "CMD=cargo +2e026d6e5d7621634147ad4c8074f1d964dbb1be-alt build --frozen" "-e" "CARGO_TARGET_DIR=/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/cargo-home" "-e" "RUSTUP_HOME=/rustup-home" "-m" "1536M" "--network" "none" "crater"` Nov 18 06:52:57.083 INFO [stdout] 1a3285d8713f568f54461c2f0a3dfe043c73c40dc1f5d449c1637fd955fc25b4 Nov 18 06:52:57.088 INFO running `"docker" "start" "-a" "1a3285d8713f568f54461c2f0a3dfe043c73c40dc1f5d449c1637fd955fc25b4"` Nov 18 06:52:58.313 INFO [stderr] usermod: no changes Nov 18 06:52:58.395 INFO [stderr] Compiling luthor v0.2.0 (/source) Nov 18 06:52:59.351 INFO [stderr] Finished dev [unoptimized + debuginfo] target(s) in 1.00s Nov 18 06:52:59.354 INFO [stderr] su: No module specific data is present Nov 18 06:52:59.668 INFO running `"docker" "inspect" "1a3285d8713f568f54461c2f0a3dfe043c73c40dc1f5d449c1637fd955fc25b4"` Nov 18 06:52:59.891 INFO running `"docker" "rm" "-f" "1a3285d8713f568f54461c2f0a3dfe043c73c40dc1f5d449c1637fd955fc25b4"` Nov 18 06:53:00.081 INFO [stdout] 1a3285d8713f568f54461c2f0a3dfe043c73c40dc1f5d449c1637fd955fc25b4 Nov 18 06:53:00.086 INFO running `"docker" "create" "-v" "/mnt/big/crater/work/local/target-dirs/pr-55982/worker-5/try#2e026d6e5d7621634147ad4c8074f1d964dbb1be:/target:rw,Z" "-v" "/mnt/big/crater/work/local/test-source/worker-5/pr-55982/try#2e026d6e5d7621634147ad4c8074f1d964dbb1be:/source:ro,Z" "-v" "/mnt/big/crater/work/local/cargo-home:/cargo-home:ro,Z" "-v" "/mnt/big/crater/work/local/rustup-home:/rustup-home:ro,Z" "-e" "USER_ID=1000" "-e" "SOURCE_DIR=/source" "-e" "USER_ID=1000" "-e" "CMD=cargo +2e026d6e5d7621634147ad4c8074f1d964dbb1be-alt test --frozen --no-run" "-e" "CARGO_TARGET_DIR=/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/cargo-home" "-e" "RUSTUP_HOME=/rustup-home" "-m" "1536M" "--network" "none" "crater"` Nov 18 06:53:00.460 INFO [stdout] 12f759c568499727fa9a8b6df48e90009ad56578b9ec2e7a1f7231066f5dc1e7 Nov 18 06:53:00.465 INFO running `"docker" "start" "-a" "12f759c568499727fa9a8b6df48e90009ad56578b9ec2e7a1f7231066f5dc1e7"` Nov 18 06:53:00.907 INFO [stderr] usermod: no changes Nov 18 06:53:00.938 INFO [stderr] Compiling luthor v0.2.0 (/source) Nov 18 06:53:02.833 INFO [stderr] Finished dev [unoptimized + debuginfo] target(s) in 1.91s Nov 18 06:53:02.836 INFO [stderr] su: No module specific data is present Nov 18 06:53:03.338 INFO running `"docker" "inspect" "12f759c568499727fa9a8b6df48e90009ad56578b9ec2e7a1f7231066f5dc1e7"` Nov 18 06:53:03.529 INFO running `"docker" "rm" "-f" "12f759c568499727fa9a8b6df48e90009ad56578b9ec2e7a1f7231066f5dc1e7"` Nov 18 06:53:03.789 INFO [stdout] 12f759c568499727fa9a8b6df48e90009ad56578b9ec2e7a1f7231066f5dc1e7 Nov 18 06:53:03.795 INFO running `"docker" "create" "-v" "/mnt/big/crater/work/local/target-dirs/pr-55982/worker-5/try#2e026d6e5d7621634147ad4c8074f1d964dbb1be:/target:rw,Z" "-v" "/mnt/big/crater/work/local/test-source/worker-5/pr-55982/try#2e026d6e5d7621634147ad4c8074f1d964dbb1be:/source:ro,Z" "-v" "/mnt/big/crater/work/local/cargo-home:/cargo-home:ro,Z" "-v" "/mnt/big/crater/work/local/rustup-home:/rustup-home:ro,Z" "-e" "USER_ID=1000" "-e" "SOURCE_DIR=/source" "-e" "USER_ID=1000" "-e" "CMD=cargo +2e026d6e5d7621634147ad4c8074f1d964dbb1be-alt test --frozen" "-e" "CARGO_TARGET_DIR=/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/cargo-home" "-e" "RUSTUP_HOME=/rustup-home" "-m" "1536M" "--network" "none" "crater"` Nov 18 06:53:04.175 INFO [stdout] 382a8636b2bf32b57026a87461b65714cf3902703391118de292e4ce4f639f1e Nov 18 06:53:04.183 INFO running `"docker" "start" "-a" "382a8636b2bf32b57026a87461b65714cf3902703391118de292e4ce4f639f1e"` Nov 18 06:53:05.011 INFO [stderr] usermod: no changes Nov 18 06:53:05.064 INFO [stderr] Finished dev [unoptimized + debuginfo] target(s) in 0.03s Nov 18 06:53:05.064 INFO [stderr] Running /target/debug/deps/luthor-f24edaf9ad829563 Nov 18 06:53:05.072 INFO [stdout] Nov 18 06:53:05.072 INFO [stdout] running 22 tests Nov 18 06:53:05.072 INFO [stdout] test lexers::json::tests::it_can_handle_utf8_data ... ok Nov 18 06:53:05.072 INFO [stdout] test lexers::json::tests::it_can_handle_garbage ... ok Nov 18 06:53:05.072 INFO [stdout] test lexers::xml::tests::it_can_handle_open_strings ... ok Nov 18 06:53:05.072 INFO [stdout] test lexers::xml::tests::it_can_handle_garbage ... ok Nov 18 06:53:05.072 INFO [stdout] test lexers::xml::tests::it_can_handle_utf8_data ... ok Nov 18 06:53:05.072 INFO [stdout] test lexers::xml::tests::it_works ... ok Nov 18 06:53:05.072 INFO [stdout] test lexers::rust::tests::it_works ... ok Nov 18 06:53:05.072 INFO [stdout] test lexers::default::tests::it_works ... ok Nov 18 06:53:05.072 INFO [stdout] test tokenizer::tests::current_char_returns_the_char_at_head ... ok Nov 18 06:53:05.072 INFO [stdout] test tokenizer::tests::current_char_returns_none_if_at_the_end ... ok Nov 18 06:53:05.072 INFO [stdout] test tokenizer::tests::consume_whitespace_handles_preexisting_noncategorized_chars ... ok Nov 18 06:53:05.072 INFO [stdout] test lexers::ruby::tests::it_works ... ok Nov 18 06:53:05.072 INFO [stdout] test tokenizer::tests::tokenize_next_tokenizes_previous_data_as_text ... ok Nov 18 06:53:05.072 INFO [stdout] test tokenizer::tests::tokenize_next_tokenizes_next_x_chars ... ok Nov 18 06:53:05.072 INFO [stdout] test tokenizer::tests::tokenize_does_nothing_if_range_is_empty ... ok Nov 18 06:53:05.072 INFO [stdout] test tokenizer::tests::tokenize_next_takes_at_most_what_is_left ... ok Nov 18 06:53:05.072 INFO [stdout] test tokenizer::tests::tokenize_creates_the_correct_token ... ok Nov 18 06:53:05.072 INFO [stdout] test tokenizer::tests::tokens_joins_advanced_data_with_unprocessed_data_as_text_token ... ok Nov 18 06:53:05.072 INFO [stdout] test tokenizer::tests::tokens_returns_unprocessed_data_as_text_token ... ok Nov 18 06:53:05.082 INFO [stdout] test lexers::json::tests::it_can_handle_open_strings ... ok Nov 18 06:53:05.082 INFO [stderr] Doc-tests luthor Nov 18 06:53:05.082 INFO [stdout] test lexers::ruby::tests::it_identifies_integers_and_operators ... ok Nov 18 06:53:05.082 INFO [stdout] test lexers::json::tests::it_works ... ok Nov 18 06:53:05.082 INFO [stdout] Nov 18 06:53:05.082 INFO [stdout] test result: ok. 22 passed; 0 failed; 0 ignored; 0 measured; 0 filtered out Nov 18 06:53:05.082 INFO [stdout] Nov 18 06:53:05.771 INFO [stdout] Nov 18 06:53:05.771 INFO [stdout] running 10 tests Nov 18 06:53:08.221 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::current_char (line 115) ... ok Nov 18 06:53:08.232 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::tokenize (line 238) ... ok Nov 18 06:53:08.433 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::consume_whitespace (line 307) ... ok Nov 18 06:53:08.623 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::has_prefix (line 163) ... ok Nov 18 06:53:08.624 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::new (line 29) ... ok Nov 18 06:53:08.849 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::next_non_whitespace_char (line 139) ... ok Nov 18 06:53:08.995 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::starts_with_lexeme (line 191) ... ok Nov 18 06:53:09.191 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::advance (line 90) ... ok Nov 18 06:53:10.077 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::tokens (line 48) ... ok Nov 18 06:53:10.082 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::tokenize_next (line 270) ... ok Nov 18 06:53:10.082 INFO [stdout] Nov 18 06:53:10.082 INFO [stdout] test result: ok. 10 passed; 0 failed; 0 ignored; 0 measured; 0 filtered out Nov 18 06:53:10.082 INFO [stdout] Nov 18 06:53:10.111 INFO [stderr] su: No module specific data is present Nov 18 06:53:11.405 INFO running `"docker" "inspect" "382a8636b2bf32b57026a87461b65714cf3902703391118de292e4ce4f639f1e"` Nov 18 06:53:11.753 INFO running `"docker" "rm" "-f" "382a8636b2bf32b57026a87461b65714cf3902703391118de292e4ce4f639f1e"` Nov 18 06:53:12.079 INFO [stdout] 382a8636b2bf32b57026a87461b65714cf3902703391118de292e4ce4f639f1e