Nov 18 06:53:00.721 INFO testing luthor-0.2.0 against master#6b9b97bd9b704f85f0184f7a213cc4d62bd9654c for pr-55982 Nov 18 06:53:00.721 INFO running `"docker" "create" "-v" "/mnt/big/crater/work/local/target-dirs/pr-55982/worker-6/master#6b9b97bd9b704f85f0184f7a213cc4d62bd9654c:/target:rw,Z" "-v" "/mnt/big/crater/work/local/test-source/worker-6/pr-55982/master#6b9b97bd9b704f85f0184f7a213cc4d62bd9654c:/source:ro,Z" "-v" "/mnt/big/crater/work/local/cargo-home:/cargo-home:ro,Z" "-v" "/mnt/big/crater/work/local/rustup-home:/rustup-home:ro,Z" "-e" "USER_ID=1000" "-e" "SOURCE_DIR=/source" "-e" "USER_ID=1000" "-e" "CMD=cargo +6b9b97bd9b704f85f0184f7a213cc4d62bd9654c-alt build --frozen" "-e" "CARGO_TARGET_DIR=/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/cargo-home" "-e" "RUSTUP_HOME=/rustup-home" "-m" "1536M" "--network" "none" "crater"` Nov 18 06:53:00.950 INFO [stdout] e949c11f782671428137a5840cad46010ac2113d82277bb265337c23231eb187 Nov 18 06:53:00.955 INFO running `"docker" "start" "-a" "e949c11f782671428137a5840cad46010ac2113d82277bb265337c23231eb187"` Nov 18 06:53:01.805 INFO [stderr] usermod: no changes Nov 18 06:53:01.851 INFO [stderr] Compiling luthor v0.2.0 (/source) Nov 18 06:53:02.983 INFO [stderr] Finished dev [unoptimized + debuginfo] target(s) in 1.15s Nov 18 06:53:02.987 INFO [stderr] su: No module specific data is present Nov 18 06:53:03.471 INFO running `"docker" "inspect" "e949c11f782671428137a5840cad46010ac2113d82277bb265337c23231eb187"` Nov 18 06:53:03.700 INFO running `"docker" "rm" "-f" "e949c11f782671428137a5840cad46010ac2113d82277bb265337c23231eb187"` Nov 18 06:53:03.903 INFO [stdout] e949c11f782671428137a5840cad46010ac2113d82277bb265337c23231eb187 Nov 18 06:53:03.908 INFO running `"docker" "create" "-v" "/mnt/big/crater/work/local/target-dirs/pr-55982/worker-6/master#6b9b97bd9b704f85f0184f7a213cc4d62bd9654c:/target:rw,Z" "-v" "/mnt/big/crater/work/local/test-source/worker-6/pr-55982/master#6b9b97bd9b704f85f0184f7a213cc4d62bd9654c:/source:ro,Z" "-v" "/mnt/big/crater/work/local/cargo-home:/cargo-home:ro,Z" "-v" "/mnt/big/crater/work/local/rustup-home:/rustup-home:ro,Z" "-e" "USER_ID=1000" "-e" "SOURCE_DIR=/source" "-e" "USER_ID=1000" "-e" "CMD=cargo +6b9b97bd9b704f85f0184f7a213cc4d62bd9654c-alt test --frozen --no-run" "-e" "CARGO_TARGET_DIR=/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/cargo-home" "-e" "RUSTUP_HOME=/rustup-home" "-m" "1536M" "--network" "none" "crater"` Nov 18 06:53:04.302 INFO [stdout] 6b2e8b0972605e02ab2864861b3fc3f6217717a095f3faed2c1fd4c29c87adc1 Nov 18 06:53:04.303 INFO running `"docker" "start" "-a" "6b2e8b0972605e02ab2864861b3fc3f6217717a095f3faed2c1fd4c29c87adc1"` Nov 18 06:53:05.123 INFO [stderr] usermod: no changes Nov 18 06:53:05.151 INFO [stderr] Compiling luthor v0.2.0 (/source) Nov 18 06:53:06.983 INFO [stderr] Finished dev [unoptimized + debuginfo] target(s) in 1.84s Nov 18 06:53:06.993 INFO [stderr] su: No module specific data is present Nov 18 06:53:08.040 INFO running `"docker" "inspect" "6b2e8b0972605e02ab2864861b3fc3f6217717a095f3faed2c1fd4c29c87adc1"` Nov 18 06:53:08.185 INFO running `"docker" "rm" "-f" "6b2e8b0972605e02ab2864861b3fc3f6217717a095f3faed2c1fd4c29c87adc1"` Nov 18 06:53:08.510 INFO [stdout] 6b2e8b0972605e02ab2864861b3fc3f6217717a095f3faed2c1fd4c29c87adc1 Nov 18 06:53:08.511 INFO running `"docker" "create" "-v" "/mnt/big/crater/work/local/target-dirs/pr-55982/worker-6/master#6b9b97bd9b704f85f0184f7a213cc4d62bd9654c:/target:rw,Z" "-v" "/mnt/big/crater/work/local/test-source/worker-6/pr-55982/master#6b9b97bd9b704f85f0184f7a213cc4d62bd9654c:/source:ro,Z" "-v" "/mnt/big/crater/work/local/cargo-home:/cargo-home:ro,Z" "-v" "/mnt/big/crater/work/local/rustup-home:/rustup-home:ro,Z" "-e" "USER_ID=1000" "-e" "SOURCE_DIR=/source" "-e" "USER_ID=1000" "-e" "CMD=cargo +6b9b97bd9b704f85f0184f7a213cc4d62bd9654c-alt test --frozen" "-e" "CARGO_TARGET_DIR=/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/cargo-home" "-e" "RUSTUP_HOME=/rustup-home" "-m" "1536M" "--network" "none" "crater"` Nov 18 06:53:08.847 INFO [stdout] 647c9de93e31238638aac71d981cf716698e5d8f0b209a26a8837f54a00bf927 Nov 18 06:53:08.849 INFO running `"docker" "start" "-a" "647c9de93e31238638aac71d981cf716698e5d8f0b209a26a8837f54a00bf927"` Nov 18 06:53:09.749 INFO [stderr] usermod: no changes Nov 18 06:53:09.799 INFO [stderr] Finished dev [unoptimized + debuginfo] target(s) in 0.03s Nov 18 06:53:09.799 INFO [stderr] Running /target/debug/deps/luthor-81f43357da2e7492 Nov 18 06:53:09.805 INFO [stdout] Nov 18 06:53:09.805 INFO [stdout] running 22 tests Nov 18 06:53:09.812 INFO [stdout] test lexers::json::tests::it_can_handle_garbage ... ok Nov 18 06:53:09.812 INFO [stdout] test lexers::default::tests::it_works ... ok Nov 18 06:53:09.812 INFO [stdout] test lexers::json::tests::it_can_handle_open_strings ... ok Nov 18 06:53:09.819 INFO [stdout] test lexers::ruby::tests::it_identifies_integers_and_operators ... ok Nov 18 06:53:09.819 INFO [stdout] test lexers::json::tests::it_can_handle_utf8_data ... ok Nov 18 06:53:09.819 INFO [stdout] test lexers::ruby::tests::it_works ... ok Nov 18 06:53:09.819 INFO [stdout] test lexers::xml::tests::it_works ... ok Nov 18 06:53:09.819 INFO [stdout] test tokenizer::tests::consume_whitespace_handles_preexisting_noncategorized_chars ... ok Nov 18 06:53:09.819 INFO [stdout] test lexers::rust::tests::it_works ... ok Nov 18 06:53:09.819 INFO [stdout] test tokenizer::tests::tokenize_does_nothing_if_range_is_empty ... ok Nov 18 06:53:09.819 INFO [stdout] test lexers::xml::tests::it_can_handle_garbage ... ok Nov 18 06:53:09.819 INFO [stdout] test tokenizer::tests::current_char_returns_none_if_at_the_end ... ok Nov 18 06:53:09.819 INFO [stdout] test lexers::xml::tests::it_can_handle_open_strings ... ok Nov 18 06:53:09.819 INFO [stdout] test lexers::xml::tests::it_can_handle_utf8_data ... ok Nov 18 06:53:09.819 INFO [stdout] test tokenizer::tests::tokenize_creates_the_correct_token ... ok Nov 18 06:53:09.819 INFO [stdout] test tokenizer::tests::tokenize_next_tokenizes_previous_data_as_text ... ok Nov 18 06:53:09.819 INFO [stdout] test tokenizer::tests::tokenize_next_takes_at_most_what_is_left ... ok Nov 18 06:53:09.819 INFO [stdout] test tokenizer::tests::current_char_returns_the_char_at_head ... ok Nov 18 06:53:09.819 INFO [stdout] test tokenizer::tests::tokenize_next_tokenizes_next_x_chars ... ok Nov 18 06:53:09.819 INFO [stdout] test tokenizer::tests::tokens_joins_advanced_data_with_unprocessed_data_as_text_token ... ok Nov 18 06:53:09.819 INFO [stdout] test tokenizer::tests::tokens_returns_unprocessed_data_as_text_token ... ok Nov 18 06:53:09.819 INFO [stdout] test lexers::json::tests::it_works ... ok Nov 18 06:53:09.819 INFO [stdout] Nov 18 06:53:09.819 INFO [stdout] test result: ok. 22 passed; 0 failed; 0 ignored; 0 measured; 0 filtered out Nov 18 06:53:09.819 INFO [stdout] Nov 18 06:53:09.823 INFO [stderr] Doc-tests luthor Nov 18 06:53:10.691 INFO [stdout] Nov 18 06:53:10.691 INFO [stdout] running 10 tests Nov 18 06:53:13.566 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::has_prefix (line 163) ... ok Nov 18 06:53:14.189 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::new (line 29) ... ok Nov 18 06:53:14.937 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::consume_whitespace (line 307) ... ok Nov 18 06:53:14.939 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::next_non_whitespace_char (line 139) ... ok Nov 18 06:53:15.150 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::starts_with_lexeme (line 191) ... ok Nov 18 06:53:15.595 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::tokenize (line 238) ... ok Nov 18 06:53:15.742 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::current_char (line 115) ... ok Nov 18 06:53:15.805 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::advance (line 90) ... ok Nov 18 06:53:16.125 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::tokenize_next (line 270) ... ok Nov 18 06:53:16.163 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::tokens (line 48) ... ok Nov 18 06:53:16.163 INFO [stdout] Nov 18 06:53:16.163 INFO [stdout] test result: ok. 10 passed; 0 failed; 0 ignored; 0 measured; 0 filtered out Nov 18 06:53:16.163 INFO [stdout] Nov 18 06:53:16.182 INFO [stderr] su: No module specific data is present Nov 18 06:53:16.512 INFO running `"docker" "inspect" "647c9de93e31238638aac71d981cf716698e5d8f0b209a26a8837f54a00bf927"` Nov 18 06:53:16.618 INFO running `"docker" "rm" "-f" "647c9de93e31238638aac71d981cf716698e5d8f0b209a26a8837f54a00bf927"` Nov 18 06:53:16.714 INFO [stdout] 647c9de93e31238638aac71d981cf716698e5d8f0b209a26a8837f54a00bf927