Dec 07 07:58:44.877 INFO testing luthor-0.2.0 against master#906deae0790bd18681b937fe9a141a3c26cf1855 for pr-56514 Dec 07 07:58:44.877 INFO running `"docker" "create" "-v" "/mnt/big/crater/work/local/target-dirs/pr-56514/worker-3/master#906deae0790bd18681b937fe9a141a3c26cf1855:/target:rw,Z" "-v" "/mnt/big/crater/work/local/test-source/worker-3/pr-56514/master#906deae0790bd18681b937fe9a141a3c26cf1855:/source:ro,Z" "-v" "/mnt/big/crater/work/local/cargo-home:/cargo-home:ro,Z" "-v" "/mnt/big/crater/work/local/rustup-home:/rustup-home:ro,Z" "-e" "USER_ID=1000" "-e" "SOURCE_DIR=/source" "-e" "USER_ID=1000" "-e" "CMD=cargo +906deae0790bd18681b937fe9a141a3c26cf1855-alt build --frozen" "-e" "CARGO_TARGET_DIR=/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/cargo-home" "-e" "RUSTUP_HOME=/rustup-home" "-m" "1536M" "--network" "none" "crater"` Dec 07 07:58:45.202 INFO [stdout] 07a8b64dcde91286303aa5e7d1653d1153a42ad5d5a1dd9ea28d1ef5c44562bf Dec 07 07:58:45.206 INFO running `"docker" "start" "-a" "07a8b64dcde91286303aa5e7d1653d1153a42ad5d5a1dd9ea28d1ef5c44562bf"` Dec 07 07:58:46.000 INFO [stderr] usermod: no changes Dec 07 07:58:46.040 INFO [stderr] Compiling luthor v0.2.0 (/source) Dec 07 07:58:47.075 INFO [stderr] Finished dev [unoptimized + debuginfo] target(s) in 1.04s Dec 07 07:58:47.075 INFO [stderr] su: No module specific data is present Dec 07 07:58:47.590 INFO running `"docker" "inspect" "07a8b64dcde91286303aa5e7d1653d1153a42ad5d5a1dd9ea28d1ef5c44562bf"` Dec 07 07:58:47.910 INFO running `"docker" "rm" "-f" "07a8b64dcde91286303aa5e7d1653d1153a42ad5d5a1dd9ea28d1ef5c44562bf"` Dec 07 07:58:48.163 INFO [stdout] 07a8b64dcde91286303aa5e7d1653d1153a42ad5d5a1dd9ea28d1ef5c44562bf Dec 07 07:58:48.166 INFO running `"docker" "create" "-v" "/mnt/big/crater/work/local/target-dirs/pr-56514/worker-3/master#906deae0790bd18681b937fe9a141a3c26cf1855:/target:rw,Z" "-v" "/mnt/big/crater/work/local/test-source/worker-3/pr-56514/master#906deae0790bd18681b937fe9a141a3c26cf1855:/source:ro,Z" "-v" "/mnt/big/crater/work/local/cargo-home:/cargo-home:ro,Z" "-v" "/mnt/big/crater/work/local/rustup-home:/rustup-home:ro,Z" "-e" "USER_ID=1000" "-e" "SOURCE_DIR=/source" "-e" "USER_ID=1000" "-e" "CMD=cargo +906deae0790bd18681b937fe9a141a3c26cf1855-alt test --frozen --no-run" "-e" "CARGO_TARGET_DIR=/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/cargo-home" "-e" "RUSTUP_HOME=/rustup-home" "-m" "1536M" "--network" "none" "crater"` Dec 07 07:58:48.457 INFO [stdout] f5edce9ee6bd154293dedd152acfcd24d9763e3103e1a5daf7dbb4d4727f0bfb Dec 07 07:58:48.462 INFO running `"docker" "start" "-a" "f5edce9ee6bd154293dedd152acfcd24d9763e3103e1a5daf7dbb4d4727f0bfb"` Dec 07 07:58:49.772 INFO [stderr] usermod: no changes Dec 07 07:58:49.852 INFO [stderr] Compiling luthor v0.2.0 (/source) Dec 07 07:58:51.805 INFO [stderr] Finished dev [unoptimized + debuginfo] target(s) in 1.99s Dec 07 07:58:51.807 INFO [stderr] su: No module specific data is present Dec 07 07:58:52.072 INFO running `"docker" "inspect" "f5edce9ee6bd154293dedd152acfcd24d9763e3103e1a5daf7dbb4d4727f0bfb"` Dec 07 07:58:52.164 INFO running `"docker" "rm" "-f" "f5edce9ee6bd154293dedd152acfcd24d9763e3103e1a5daf7dbb4d4727f0bfb"` Dec 07 07:58:52.287 INFO [stdout] f5edce9ee6bd154293dedd152acfcd24d9763e3103e1a5daf7dbb4d4727f0bfb Dec 07 07:58:52.290 INFO running `"docker" "create" "-v" "/mnt/big/crater/work/local/target-dirs/pr-56514/worker-3/master#906deae0790bd18681b937fe9a141a3c26cf1855:/target:rw,Z" "-v" "/mnt/big/crater/work/local/test-source/worker-3/pr-56514/master#906deae0790bd18681b937fe9a141a3c26cf1855:/source:ro,Z" "-v" "/mnt/big/crater/work/local/cargo-home:/cargo-home:ro,Z" "-v" "/mnt/big/crater/work/local/rustup-home:/rustup-home:ro,Z" "-e" "USER_ID=1000" "-e" "SOURCE_DIR=/source" "-e" "USER_ID=1000" "-e" "CMD=cargo +906deae0790bd18681b937fe9a141a3c26cf1855-alt test --frozen" "-e" "CARGO_TARGET_DIR=/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/cargo-home" "-e" "RUSTUP_HOME=/rustup-home" "-m" "1536M" "--network" "none" "crater"` Dec 07 07:58:52.544 INFO [stdout] 7beb2d050a36cf1984cf6ccd619bac09bdbe214d5c097d096b50c1caa24ceac1 Dec 07 07:58:52.546 INFO running `"docker" "start" "-a" "7beb2d050a36cf1984cf6ccd619bac09bdbe214d5c097d096b50c1caa24ceac1"` Dec 07 07:58:52.937 INFO [stderr] usermod: no changes Dec 07 07:58:52.969 INFO [stderr] Finished dev [unoptimized + debuginfo] target(s) in 0.01s Dec 07 07:58:52.973 INFO [stderr] Running /target/debug/deps/luthor-88162204a99e25a2 Dec 07 07:58:52.974 INFO [stdout] Dec 07 07:58:52.974 INFO [stdout] running 22 tests Dec 07 07:58:52.975 INFO [stdout] test lexers::json::tests::it_can_handle_utf8_data ... ok Dec 07 07:58:52.975 INFO [stdout] test lexers::ruby::tests::it_identifies_integers_and_operators ... ok Dec 07 07:58:52.975 INFO [stdout] test lexers::json::tests::it_can_handle_garbage ... ok Dec 07 07:58:52.975 INFO [stdout] test lexers::default::tests::it_works ... ok Dec 07 07:58:52.976 INFO [stdout] test lexers::json::tests::it_can_handle_open_strings ... ok Dec 07 07:58:52.976 INFO [stdout] test lexers::json::tests::it_works ... ok Dec 07 07:58:52.976 INFO [stdout] test lexers::ruby::tests::it_works ... ok Dec 07 07:58:52.976 INFO [stdout] test lexers::xml::tests::it_can_handle_open_strings ... ok Dec 07 07:58:52.976 INFO [stdout] test tokenizer::tests::tokenize_creates_the_correct_token ... ok Dec 07 07:58:52.976 INFO [stdout] test lexers::xml::tests::it_can_handle_utf8_data ... ok Dec 07 07:58:52.976 INFO [stdout] test lexers::xml::tests::it_works ... ok Dec 07 07:58:52.976 INFO [stdout] test tokenizer::tests::consume_whitespace_handles_preexisting_noncategorized_chars ... ok Dec 07 07:58:52.976 INFO [stdout] test tokenizer::tests::tokenize_next_tokenizes_next_x_chars ... ok Dec 07 07:58:52.976 INFO [stdout] test tokenizer::tests::tokens_joins_advanced_data_with_unprocessed_data_as_text_token ... ok Dec 07 07:58:52.976 INFO [stdout] test tokenizer::tests::tokenize_next_takes_at_most_what_is_left ... ok Dec 07 07:58:52.976 INFO [stdout] test tokenizer::tests::tokenize_does_nothing_if_range_is_empty ... ok Dec 07 07:58:52.976 INFO [stdout] test tokenizer::tests::current_char_returns_none_if_at_the_end ... ok Dec 07 07:58:52.980 INFO [stdout] test tokenizer::tests::current_char_returns_the_char_at_head ... ok Dec 07 07:58:52.980 INFO [stdout] test lexers::rust::tests::it_works ... ok Dec 07 07:58:52.980 INFO [stdout] test lexers::xml::tests::it_can_handle_garbage ... ok Dec 07 07:58:52.980 INFO [stdout] test tokenizer::tests::tokenize_next_tokenizes_previous_data_as_text ... ok Dec 07 07:58:52.980 INFO [stdout] test tokenizer::tests::tokens_returns_unprocessed_data_as_text_token ... ok Dec 07 07:58:52.980 INFO [stdout] Dec 07 07:58:52.980 INFO [stdout] test result: ok. 22 passed; 0 failed; 0 ignored; 0 measured; 0 filtered out Dec 07 07:58:52.980 INFO [stdout] Dec 07 07:58:52.983 INFO [stderr] Doc-tests luthor Dec 07 07:58:53.688 INFO [stdout] Dec 07 07:58:53.688 INFO [stdout] running 10 tests Dec 07 07:58:58.308 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::starts_with_lexeme (line 191) ... ok Dec 07 07:58:58.316 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::current_char (line 115) ... ok Dec 07 07:58:59.453 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::advance (line 90) ... ok Dec 07 07:58:59.612 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::tokenize (line 238) ... ok Dec 07 07:58:59.612 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::next_non_whitespace_char (line 139) ... ok Dec 07 07:58:59.776 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::new (line 29) ... ok Dec 07 07:58:59.815 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::has_prefix (line 163) ... ok Dec 07 07:58:59.824 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::consume_whitespace (line 307) ... ok Dec 07 07:59:01.088 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::tokenize_next (line 270) ... ok Dec 07 07:59:01.092 INFO [stdout] test src/tokenizer.rs - tokenizer::Tokenizer<'a>::tokens (line 48) ... ok Dec 07 07:59:01.092 INFO [stdout] Dec 07 07:59:01.092 INFO [stdout] test result: ok. 10 passed; 0 failed; 0 ignored; 0 measured; 0 filtered out Dec 07 07:59:01.092 INFO [stdout] Dec 07 07:59:01.115 INFO [stderr] su: No module specific data is present Dec 07 07:59:01.634 INFO running `"docker" "inspect" "7beb2d050a36cf1984cf6ccd619bac09bdbe214d5c097d096b50c1caa24ceac1"` Dec 07 07:59:01.866 INFO running `"docker" "rm" "-f" "7beb2d050a36cf1984cf6ccd619bac09bdbe214d5c097d096b50c1caa24ceac1"` Dec 07 07:59:02.074 INFO [stdout] 7beb2d050a36cf1984cf6ccd619bac09bdbe214d5c097d096b50c1caa24ceac1