[INFO] cloning repository https://github.com/drocco007/tr [INFO] running `Command { std: "git" "-c" "credential.helper=" "-c" "credential.helper=/workspace/cargo-home/bin/git-credential-null" "clone" "--bare" "https://github.com/drocco007/tr" "/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2Fdrocco007%2Ftr", kill_on_drop: false }` [INFO] [stderr] Cloning into bare repository '/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2Fdrocco007%2Ftr'... [INFO] running `Command { std: "git" "rev-parse" "HEAD", kill_on_drop: false }` [INFO] [stdout] bf2e32c14c89341784931a20a4f90bf3b1fe8d7b [INFO] testing drocco007/tr against try#8de4c7234dd9b97c9d76b58671343fdbbc9a433e+target=x86_64-unknown-linux-musl for musl_upgrade_1_2_5_with_libc_patch_0 [INFO] running `Command { std: "git" "clone" "/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2Fdrocco007%2Ftr" "/workspace/builds/worker-4-tc1/source", kill_on_drop: false }` [INFO] [stderr] Cloning into '/workspace/builds/worker-4-tc1/source'... [INFO] [stderr] done. [INFO] started tweaking git repo https://github.com/drocco007/tr [INFO] finished tweaking git repo https://github.com/drocco007/tr [INFO] tweaked toml for git repo https://github.com/drocco007/tr written to /workspace/builds/worker-4-tc1/source/Cargo.toml [INFO] validating manifest of git repo https://github.com/drocco007/tr on toolchain 8de4c7234dd9b97c9d76b58671343fdbbc9a433e [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+8de4c7234dd9b97c9d76b58671343fdbbc9a433e" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }` [INFO] crate git repo https://github.com/drocco007/tr already has a lockfile, it will not be regenerated [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+8de4c7234dd9b97c9d76b58671343fdbbc9a433e" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }` [INFO] [stderr] Updating crates.io index [INFO] [stderr] Downloading crates ... [INFO] [stderr] Downloaded syn v1.0.46 [INFO] [stderr] Downloaded rstest v0.6.4 [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-4-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-4-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:90999bfc7ae267e83380e433d8e61a7c072ca6729e92edbae886d3423b3a6f4c" "/opt/rustwide/cargo-home/bin/cargo" "+8de4c7234dd9b97c9d76b58671343fdbbc9a433e" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }` [INFO] [stdout] 0e729a126b7f1bbada5981ad3c9c819f800f5deaba113116a08cf3b1b9d8c602 [INFO] running `Command { std: "docker" "start" "-a" "0e729a126b7f1bbada5981ad3c9c819f800f5deaba113116a08cf3b1b9d8c602", kill_on_drop: false }` [INFO] running `Command { std: "docker" "inspect" "0e729a126b7f1bbada5981ad3c9c819f800f5deaba113116a08cf3b1b9d8c602", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "0e729a126b7f1bbada5981ad3c9c819f800f5deaba113116a08cf3b1b9d8c602", kill_on_drop: false }` [INFO] [stdout] 0e729a126b7f1bbada5981ad3c9c819f800f5deaba113116a08cf3b1b9d8c602 [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-4-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-4-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:90999bfc7ae267e83380e433d8e61a7c072ca6729e92edbae886d3423b3a6f4c" "/opt/rustwide/cargo-home/bin/cargo" "+8de4c7234dd9b97c9d76b58671343fdbbc9a433e" "build" "--frozen" "--message-format=json" "--target" "x86_64-unknown-linux-musl", kill_on_drop: false }` [INFO] [stdout] 420ffb9ecf23b42fd827a7fadb3cc9f362dfff6ccafb0becf9200826baf4682a [INFO] running `Command { std: "docker" "start" "-a" "420ffb9ecf23b42fd827a7fadb3cc9f362dfff6ccafb0becf9200826baf4682a", kill_on_drop: false }` [INFO] [stderr] Compiling byteorder v1.3.4 [INFO] [stderr] Compiling memchr v2.3.4 [INFO] [stderr] Compiling regex-automata v0.1.9 [INFO] [stderr] Compiling bstr v0.2.14 [INFO] [stderr] Compiling tr v0.1.0 (/opt/rustwide/workdir) [INFO] [stdout] warning: unnecessary trailing semicolon [INFO] [stdout] --> src/lex.rs:81:10 [INFO] [stdout] | [INFO] [stdout] 81 | }; [INFO] [stdout] | ^ help: remove this semicolon [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(redundant_semicolons)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused `Result` that must be used [INFO] [stdout] --> src/command.rs:191:9 [INFO] [stdout] | [INFO] [stdout] 191 | tr.process(); [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: this `Result` may be an `Err` variant, which should be handled [INFO] [stdout] = note: `#[warn(unused_must_use)]` on by default [INFO] [stdout] help: use `let _ = ...` to ignore the resulting value [INFO] [stdout] | [INFO] [stdout] 191 | let _ = tr.process(); [INFO] [stdout] | +++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: lifetime flowing from input to output with different syntax can be confusing [INFO] [stdout] --> src/lex.rs:273:20 [INFO] [stdout] | [INFO] [stdout] 273 | pub fn tokenize(s: &str) -> Lexer { [INFO] [stdout] | ^^^^ ----- the lifetime gets resolved as `'_` [INFO] [stdout] | | [INFO] [stdout] | this lifetime flows to the output [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(mismatched_lifetime_syntaxes)]` on by default [INFO] [stdout] help: one option is to remove the lifetime for references and use the anonymous lifetime for paths [INFO] [stdout] | [INFO] [stdout] 273 | pub fn tokenize(s: &str) -> Lexer<'_> { [INFO] [stdout] | ++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] Finished `dev` profile [unoptimized + debuginfo] target(s) in 2.83s [INFO] running `Command { std: "docker" "inspect" "420ffb9ecf23b42fd827a7fadb3cc9f362dfff6ccafb0becf9200826baf4682a", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "420ffb9ecf23b42fd827a7fadb3cc9f362dfff6ccafb0becf9200826baf4682a", kill_on_drop: false }` [INFO] [stdout] 420ffb9ecf23b42fd827a7fadb3cc9f362dfff6ccafb0becf9200826baf4682a [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-4-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-4-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:90999bfc7ae267e83380e433d8e61a7c072ca6729e92edbae886d3423b3a6f4c" "/opt/rustwide/cargo-home/bin/cargo" "+8de4c7234dd9b97c9d76b58671343fdbbc9a433e" "test" "--frozen" "--no-run" "--message-format=json" "--target" "x86_64-unknown-linux-musl", kill_on_drop: false }` [INFO] [stdout] 88301e5cf7273f6dfc6387adab56200cd16de643d6fdf6f6abc96d77497c9192 [INFO] running `Command { std: "docker" "start" "-a" "88301e5cf7273f6dfc6387adab56200cd16de643d6fdf6f6abc96d77497c9192", kill_on_drop: false }` [INFO] [stderr] Compiling proc-macro2 v1.0.24 [INFO] [stderr] Compiling semver-parser v0.7.0 [INFO] [stderr] Compiling unicode-xid v0.2.1 [INFO] [stderr] Compiling syn v1.0.46 [INFO] [stderr] Compiling cfg-if v0.1.10 [INFO] [stdout] warning: unnecessary trailing semicolon [INFO] [stdout] --> src/lex.rs:81:10 [INFO] [stdout] | [INFO] [stdout] 81 | }; [INFO] [stdout] | ^ help: remove this semicolon [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(redundant_semicolons)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused `Result` that must be used [INFO] [stdout] --> src/command.rs:191:9 [INFO] [stdout] | [INFO] [stdout] 191 | tr.process(); [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: this `Result` may be an `Err` variant, which should be handled [INFO] [stdout] = note: `#[warn(unused_must_use)]` on by default [INFO] [stdout] help: use `let _ = ...` to ignore the resulting value [INFO] [stdout] | [INFO] [stdout] 191 | let _ = tr.process(); [INFO] [stdout] | +++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: lifetime flowing from input to output with different syntax can be confusing [INFO] [stdout] --> src/lex.rs:273:20 [INFO] [stdout] | [INFO] [stdout] 273 | pub fn tokenize(s: &str) -> Lexer { [INFO] [stdout] | ^^^^ ----- the lifetime gets resolved as `'_` [INFO] [stdout] | | [INFO] [stdout] | this lifetime flows to the output [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(mismatched_lifetime_syntaxes)]` on by default [INFO] [stdout] help: one option is to remove the lifetime for references and use the anonymous lifetime for paths [INFO] [stdout] | [INFO] [stdout] 273 | pub fn tokenize(s: &str) -> Lexer<'_> { [INFO] [stdout] | ++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] Compiling semver v0.9.0 [INFO] [stderr] Compiling rustc_version v0.2.3 [INFO] [stderr] Compiling rstest v0.6.4 [INFO] [stderr] Compiling quote v1.0.7 [INFO] [stderr] Compiling tr v0.1.0 (/opt/rustwide/workdir) [INFO] [stdout] warning: unnecessary trailing semicolon [INFO] [stdout] --> src/lex.rs:81:10 [INFO] [stdout] | [INFO] [stdout] 81 | }; [INFO] [stdout] | ^ help: remove this semicolon [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(redundant_semicolons)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused `Result` that must be used [INFO] [stdout] --> tests/black_box.rs:15:5 [INFO] [stdout] | [INFO] [stdout] 15 | tr(args, stdin, &mut stdout); [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: this `Result` may be an `Err` variant, which should be handled [INFO] [stdout] = note: `#[warn(unused_must_use)]` on by default [INFO] [stdout] help: use `let _ = ...` to ignore the resulting value [INFO] [stdout] | [INFO] [stdout] 15 | let _ = tr(args, stdin, &mut stdout); [INFO] [stdout] | +++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused `Result` that must be used [INFO] [stdout] --> src/command.rs:191:9 [INFO] [stdout] | [INFO] [stdout] 191 | tr.process(); [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: this `Result` may be an `Err` variant, which should be handled [INFO] [stdout] = note: `#[warn(unused_must_use)]` on by default [INFO] [stdout] help: use `let _ = ...` to ignore the resulting value [INFO] [stdout] | [INFO] [stdout] 191 | let _ = tr.process(); [INFO] [stdout] | +++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: lifetime flowing from input to output with different syntax can be confusing [INFO] [stdout] --> src/lex.rs:273:20 [INFO] [stdout] | [INFO] [stdout] 273 | pub fn tokenize(s: &str) -> Lexer { [INFO] [stdout] | ^^^^ ----- the lifetime gets resolved as `'_` [INFO] [stdout] | | [INFO] [stdout] | this lifetime flows to the output [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(mismatched_lifetime_syntaxes)]` on by default [INFO] [stdout] help: one option is to remove the lifetime for references and use the anonymous lifetime for paths [INFO] [stdout] | [INFO] [stdout] 273 | pub fn tokenize(s: &str) -> Lexer<'_> { [INFO] [stdout] | ++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] Finished `test` profile [unoptimized + debuginfo] target(s) in 11.24s [INFO] running `Command { std: "docker" "inspect" "88301e5cf7273f6dfc6387adab56200cd16de643d6fdf6f6abc96d77497c9192", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "88301e5cf7273f6dfc6387adab56200cd16de643d6fdf6f6abc96d77497c9192", kill_on_drop: false }` [INFO] [stdout] 88301e5cf7273f6dfc6387adab56200cd16de643d6fdf6f6abc96d77497c9192 [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-4-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-4-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:90999bfc7ae267e83380e433d8e61a7c072ca6729e92edbae886d3423b3a6f4c" "/opt/rustwide/cargo-home/bin/cargo" "+8de4c7234dd9b97c9d76b58671343fdbbc9a433e" "test" "--frozen" "--target" "x86_64-unknown-linux-musl", kill_on_drop: false }` [INFO] [stdout] 11a066a041df87198b815c2f5ee75436cf200c6010baa11d920d65c48c23fe4b [INFO] running `Command { std: "docker" "start" "-a" "11a066a041df87198b815c2f5ee75436cf200c6010baa11d920d65c48c23fe4b", kill_on_drop: false }` [INFO] [stderr] warning: unnecessary trailing semicolon [INFO] [stderr] --> src/lex.rs:81:10 [INFO] [stderr] | [INFO] [stderr] 81 | }; [INFO] [stderr] | ^ help: remove this semicolon [INFO] [stderr] | [INFO] [stderr] = note: `#[warn(redundant_semicolons)]` on by default [INFO] [stderr] [INFO] [stderr] warning: unused `Result` that must be used [INFO] [stderr] --> src/command.rs:191:9 [INFO] [stderr] | [INFO] [stderr] 191 | tr.process(); [INFO] [stderr] | ^^^^^^^^^^^^ [INFO] [stderr] | [INFO] [stderr] = note: this `Result` may be an `Err` variant, which should be handled [INFO] [stderr] = note: `#[warn(unused_must_use)]` on by default [INFO] [stderr] help: use `let _ = ...` to ignore the resulting value [INFO] [stderr] | [INFO] [stderr] 191 | let _ = tr.process(); [INFO] [stderr] | +++++++ [INFO] [stderr] [INFO] [stderr] warning: lifetime flowing from input to output with different syntax can be confusing [INFO] [stderr] --> src/lex.rs:273:20 [INFO] [stderr] | [INFO] [stderr] 273 | pub fn tokenize(s: &str) -> Lexer { [INFO] [stderr] | ^^^^ ----- the lifetime gets resolved as `'_` [INFO] [stderr] | | [INFO] [stderr] | this lifetime flows to the output [INFO] [stderr] | [INFO] [stderr] = note: `#[warn(mismatched_lifetime_syntaxes)]` on by default [INFO] [stderr] help: one option is to remove the lifetime for references and use the anonymous lifetime for paths [INFO] [stderr] | [INFO] [stderr] 273 | pub fn tokenize(s: &str) -> Lexer<'_> { [INFO] [stderr] | ++++ [INFO] [stderr] [INFO] [stderr] warning: `tr` (lib) generated 3 warnings (run `cargo fix --lib -p tr` to apply 1 suggestion) [INFO] [stderr] warning: unused `Result` that must be used [INFO] [stderr] --> tests/black_box.rs:15:5 [INFO] [stderr] | [INFO] [stderr] 15 | tr(args, stdin, &mut stdout); [INFO] [stderr] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stderr] | [INFO] [stderr] = note: this `Result` may be an `Err` variant, which should be handled [INFO] [stderr] = note: `#[warn(unused_must_use)]` on by default [INFO] [stderr] help: use `let _ = ...` to ignore the resulting value [INFO] [stderr] | [INFO] [stderr] 15 | let _ = tr(args, stdin, &mut stdout); [INFO] [stderr] | +++++++ [INFO] [stderr] [INFO] [stderr] warning: `tr` (lib test) generated 3 warnings (3 duplicates) [INFO] [stderr] warning: `tr` (test "black_box") generated 1 warning [INFO] [stderr] Finished `test` profile [unoptimized + debuginfo] target(s) in 0.07s [INFO] [stderr] Running unittests src/lib.rs (/opt/rustwide/target/x86_64-unknown-linux-musl/debug/deps/tr-0b8cc2388e3d2d6f) [INFO] [stdout] [INFO] [stdout] running 0 tests [INFO] [stderr] Running unittests src/main.rs (/opt/rustwide/target/x86_64-unknown-linux-musl/debug/deps/tr-7f76b4dd99d0485c) [INFO] [stdout] [INFO] [stdout] test result: ok. 0 passed; 0 failed; 0 ignored; 0 measured; 0 filtered out; finished in 0.00s [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] running 0 tests [INFO] [stdout] [INFO] [stdout] test result: ok. 0 passed; 0 failed; 0 ignored; 0 measured; 0 filtered out; finished in 0.00s [INFO] [stdout] [INFO] [stderr] Running tests/arg_parser.rs (/opt/rustwide/target/x86_64-unknown-linux-musl/debug/deps/arg_parser-4193a3ac9c9e5d20) [INFO] [stdout] [INFO] [stdout] running 17 tests [INFO] [stdout] test attempting_to_load_three_sets_should_produce_error ... ok [INFO] [stdout] test double_dash_should_end_option_processing ... ok [INFO] [stdout] test redundant_switches_and_options_should_be_allowed ... ok [INFO] [stdout] test pseudo_switch_as_set1_should_indicate_unrecognized_option ... ok [INFO] [stdout] test simple_two_arguments_should_load_set1_and_set2 ... ok [INFO] [stdout] test redundant_switch_should_be_allowed ... ok [INFO] [stdout] test single_set_with_squeeze_should_be_valid ... ok [INFO] [stdout] test combined_squeeze_and_delete_without_set2_should_be_invalid ... ok [INFO] [stdout] test missing_set1_should_produce_error ... ok [INFO] [stdout] test combined_squeeze_and_delete_with_set2_should_be_valid ... ok [INFO] [stdout] test multilple_switches_specified_as_two_arguments_should_be_allowed ... ok [INFO] [stdout] test pseudo_switch_as_set1_should_indicate_invalid_option ... ok [INFO] [stdout] test single_set_with_delete_should_be_valid ... ok [INFO] [stdout] test single_set_without_squeeze_or_delete_should_be_invalid ... ok [INFO] [stdout] test unknown_switch_in_combo_should_indicate_correct_invalid_option ... ok [INFO] [stdout] test multilple_switches_specified_as_one_argument_should_be_allowed ... ok [INFO] [stdout] test solo_dash_as_set1_should_be_valid ... ok [INFO] [stdout] [INFO] [stdout] test result: ok. 17 passed; 0 failed; 0 ignored; 0 measured; 0 filtered out; finished in 0.02s [INFO] [stdout] [INFO] [stderr] Running tests/black_box.rs (/opt/rustwide/target/x86_64-unknown-linux-musl/debug/deps/black_box-512a910d8ab74ad0) [INFO] [stdout] [INFO] [stdout] running 20 tests [INFO] [stdout] test delete_should_remove_all_occurrences ... ok [INFO] [stdout] test should_delete_nothing ... ok [INFO] [stdout] test delete_should_remove_all_occurrences_of_all_set1 ... ok [INFO] [stdout] test delete_should_remove_entire_input ... ok [INFO] [stdout] test should_pad_last_of_set2_to_length_of_set1 ... ignored [INFO] [stdout] test should_pad_set2_with_interior_repeat ... ignored [INFO] [stdout] test should_pad_set2_with_mixed_explicit_and_implicit_repeat ... ignored [INFO] [stdout] test should_pad_set2_with_repeat_length_specified_in_octal ... ignored [INFO] [stdout] test should_perform_complement_delete ... ok [INFO] [stdout] test should_perform_complement_squeeze ... ok [INFO] [stdout] test should_perform_simple_delete ... ok [INFO] [stdout] test should_expand_simple_repeat ... ok [INFO] [stdout] test should_say_hello_loudly ... ok [INFO] [stdout] test simple_squeeze_should ... ok [INFO] [stdout] test simple_squeeze_should_squeeze_multiple ... ok [INFO] [stdout] test squeeze_delete_should_remove_then_squeeze ... ok [INFO] [stdout] test should_translate_plain_suit_to_fancy ... ok [INFO] [stdout] test simple_squeeze_should_preserve_multiple_not_in_set ... ok [INFO] [stdout] test simple_squeeze_should_squeeze_all_occurrences_of_set1 ... ok [INFO] [stderr] Running tests/lex.rs (/opt/rustwide/target/x86_64-unknown-linux-musl/debug/deps/lex-417df7856c0aa36b) [INFO] [stdout] test should_squeeze_newlines ... ok [INFO] [stdout] [INFO] [stdout] test result: ok. 16 passed; 0 failed; 4 ignored; 0 measured; 0 filtered out; finished in 0.03s [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] running 300 tests [INFO] [stdout] test escape_sequence_should_be_valid_equivalence_char ... ignored [INFO] [stdout] test dangling_open_bracket_followed_by_legitimate_range ... ok [INFO] [stdout] test complicated_octal_parsing_scenario ... ok [INFO] [stdout] test final_range_should_have_type_range ... ok [INFO] [stdout] test final_range_should_tokenize_togther ... ok [INFO] [stdout] test initial_backslash_escape_should_be_of_type_literal::s_1 ... ok [INFO] [stdout] test initial_backslash_escape_should_be_of_type_literal::s_2 ... ok [INFO] [stdout] test initial_backslash_escape_should_be_of_type_literal::s_3 ... ok [INFO] [stdout] test initial_backslash_escape_should_be_of_type_literal::s_4 ... ok [INFO] [stdout] test initial_backslash_escape_should_be_of_type_literal::s_5 ... ok [INFO] [stdout] test complicated_scenario ... ok [INFO] [stdout] test initial_backslash_escape_should_be_of_type_literal::s_6 ... ok [INFO] [stdout] test initial_backslash_escape_should_be_of_type_literal::s_8 ... ok [INFO] [stdout] test initial_backslash_escape_should_be_of_type_literal::s_7 ... ok [INFO] [stdout] test initial_backslash_escape_should_end_token_stream::s_1 ... ok [INFO] [stdout] test initial_backslash_escape_should_end_token_stream::s_2 ... ok [INFO] [stdout] test initial_backslash_escape_should_end_token_stream::s_3 ... ok [INFO] [stdout] test initial_backslash_escape_should_end_token_stream::s_4 ... ok [INFO] [stdout] test initial_backslash_escape_should_end_token_stream::s_5 ... ok [INFO] [stdout] test initial_backslash_escape_should_end_token_stream::s_6 ... ok [INFO] [stdout] test initial_backslash_escape_should_end_token_stream::s_7 ... ok [INFO] [stdout] test initial_backslash_escape_should_end_token_stream::s_8 ... ok [INFO] [stdout] test interior_range_should_tokenize_togther ... ok [INFO] [stdout] test octal_escape_sequence_should_be_valid_equivalence_char ... ignored [INFO] [stdout] test pseudo_repeats_and_classes_should_be_treated_as_literals::s_2 ... ok [INFO] [stdout] test pseudo_repeats_and_classes_should_be_treated_as_literals::s_3 ... ok [INFO] [stdout] test inital_range_should_tokenize_togther ... ok [INFO] [stdout] test pseudo_repeats_and_classes_should_be_treated_as_literals::s_4 ... ok [INFO] [stdout] test pseudo_repeats_and_classes_should_be_treated_as_literals::s_5 ... ok [INFO] [stdout] test pseudo_repeats_and_classes_should_be_treated_as_literals::s_6 ... ok [INFO] [stdout] test pseudo_repeats_and_classes_should_be_treated_as_literals::s_1 ... ok [INFO] [stdout] test should_tokenize_class::s_01::prefix_1::suffix_2 ... ok [INFO] [stdout] test pseudo_repeats_and_classes_should_be_treated_as_literals::s_7 ... ok [INFO] [stdout] test should_tokenize_class::s_01::prefix_1::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_01::prefix_2::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_01::prefix_2::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_01::prefix_1::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_01::prefix_2::suffix_3 ... ok [INFO] [stdout] test pseudo_repeats_and_classes_should_be_treated_as_literals::s_8 ... ok [INFO] [stdout] test should_tokenize_class::s_01::prefix_3::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_02::prefix_1::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_02::prefix_1::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_02::prefix_1::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_02::prefix_2::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_02::prefix_2::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_02::prefix_2::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_02::prefix_3::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_02::prefix_3::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_02::prefix_3::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_01::prefix_3::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_03::prefix_1::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_03::prefix_1::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_03::prefix_1::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_03::prefix_2::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_01::prefix_3::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_03::prefix_2::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_03::prefix_2::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_03::prefix_3::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_04::prefix_1::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_04::prefix_1::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_04::prefix_1::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_03::prefix_3::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_04::prefix_2::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_04::prefix_2::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_03::prefix_3::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_04::prefix_2::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_04::prefix_3::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_04::prefix_3::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_04::prefix_3::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_05::prefix_1::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_05::prefix_1::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_05::prefix_1::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_05::prefix_2::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_05::prefix_2::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_05::prefix_3::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_05::prefix_3::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_05::prefix_3::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_06::prefix_1::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_06::prefix_1::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_06::prefix_1::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_06::prefix_2::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_06::prefix_2::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_06::prefix_2::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_06::prefix_3::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_06::prefix_3::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_06::prefix_3::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_07::prefix_1::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_07::prefix_1::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_05::prefix_2::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_07::prefix_1::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_07::prefix_2::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_07::prefix_2::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_07::prefix_2::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_07::prefix_3::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_07::prefix_3::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_07::prefix_3::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_08::prefix_1::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_08::prefix_1::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_08::prefix_1::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_08::prefix_2::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_08::prefix_2::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_08::prefix_3::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_08::prefix_3::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_08::prefix_3::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_09::prefix_1::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_09::prefix_1::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_09::prefix_1::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_09::prefix_2::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_09::prefix_2::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_09::prefix_2::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_09::prefix_3::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_08::prefix_2::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_09::prefix_3::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_09::prefix_3::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_10::prefix_1::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_10::prefix_1::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_10::prefix_1::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_10::prefix_2::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_10::prefix_2::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_10::prefix_2::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_10::prefix_3::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_10::prefix_3::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_10::prefix_3::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_11::prefix_1::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_11::prefix_1::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_11::prefix_1::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_11::prefix_2::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_11::prefix_2::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_11::prefix_2::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_11::prefix_3::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_11::prefix_3::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_11::prefix_3::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_12::prefix_1::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_12::prefix_1::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_12::prefix_1::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_12::prefix_2::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_12::prefix_2::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_12::prefix_3::suffix_1 ... ok [INFO] [stdout] test should_tokenize_class::s_12::prefix_3::suffix_2 ... ok [INFO] [stdout] test should_tokenize_class::s_12::prefix_3::suffix_3 ... ok [INFO] [stdout] test should_tokenize_class::s_12::prefix_2::suffix_3 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_1::prefix_1::suffix_3 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_1::prefix_2::suffix_1 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_1::prefix_1::suffix_1 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_1::prefix_1::suffix_2 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_1::prefix_3::suffix_1 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_1::prefix_3::suffix_2 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_1::prefix_3::suffix_3 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_2::prefix_1::suffix_1 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_2::prefix_1::suffix_2 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_2::prefix_1::suffix_3 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_2::prefix_2::suffix_2 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_2::prefix_3::suffix_3 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_3::prefix_1::suffix_1 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_3::prefix_1::suffix_3 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_3::prefix_2::suffix_2 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_1::prefix_2::suffix_2 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_1::prefix_2::suffix_3 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_2::prefix_2::suffix_1 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_3::prefix_1::suffix_2 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_2::prefix_3::suffix_1 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_2::prefix_2::suffix_3 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_3::prefix_2::suffix_1 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_3::prefix_3::suffix_2 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_4::prefix_1::suffix_1 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_4::prefix_1::suffix_3 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_4::prefix_2::suffix_2 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_4::prefix_3::suffix_2 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_4::prefix_1::suffix_2 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_2::prefix_3::suffix_2 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_4::prefix_3::suffix_3 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_3::prefix_2::suffix_3 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_4::prefix_2::suffix_1 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_4::prefix_2::suffix_3 ... ok [INFO] [stdout] test should_tokenize_octal_escape::case_1::prefix_1::suffix_1 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_1::prefix_1::suffix_2 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_1::prefix_1::suffix_3 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_1::prefix_2::suffix_1 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_1::prefix_2::suffix_2 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_1::prefix_2::suffix_3 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_1::prefix_3::suffix_1 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_1::prefix_3::suffix_2 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_1::prefix_3::suffix_3 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_2::prefix_1::suffix_1 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_2::prefix_1::suffix_2 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_2::prefix_1::suffix_3 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_2::prefix_2::suffix_1 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_2::prefix_2::suffix_2 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_2::prefix_2::suffix_3 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_2::prefix_3::suffix_1 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_2::prefix_3::suffix_2 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_2::prefix_3::suffix_3 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_3::prefix_1::suffix_1 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_3::prefix_1::suffix_2 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_3::prefix_1::suffix_3 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_3::prefix_2::suffix_1 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_3::prefix_2::suffix_2 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_3::prefix_2::suffix_3 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_3::prefix_3::suffix_1 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_3::prefix_3::suffix_2 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_3::prefix_3::suffix_3 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_4::prefix_1::suffix_1 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_4::prefix_1::suffix_2 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_4::prefix_1::suffix_3 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_4::prefix_2::suffix_1 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_4::prefix_2::suffix_2 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_4::prefix_2::suffix_3 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_4::prefix_3::suffix_1 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_4::prefix_3::suffix_2 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_4::prefix_3::suffix_3 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_5::prefix_1::suffix_1 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_5::prefix_1::suffix_2 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_5::prefix_1::suffix_3 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_5::prefix_2::suffix_1 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_5::prefix_2::suffix_2 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_5::prefix_2::suffix_3 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_5::prefix_3::suffix_1 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_5::prefix_3::suffix_2 ... ignored [INFO] [stdout] test should_tokenize_octal_escape::case_5::prefix_3::suffix_3 ... ignored [INFO] [stdout] test should_tokenize_repeat::s_1::prefix_1::suffix_1 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_3::prefix_3::suffix_1 ... ok [INFO] [stdout] test should_tokenize_repeat::s_1::prefix_1::suffix_2 ... ok [INFO] [stdout] test should_tokenize_repeat::s_1::prefix_1::suffix_3 ... ok [INFO] [stdout] test should_tokenize_repeat::s_1::prefix_2::suffix_1 ... ok [INFO] [stdout] test should_tokenize_repeat::s_1::prefix_2::suffix_2 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_3::prefix_3::suffix_3 ... ok [INFO] [stdout] test should_tokenize_equivalence::s_4::prefix_3::suffix_1 ... ok [INFO] [stdout] test should_tokenize_repeat::s_1::prefix_2::suffix_3 ... ok [INFO] [stdout] test should_tokenize_repeat::s_1::prefix_3::suffix_1 ... ok [INFO] [stdout] test should_tokenize_repeat::s_1::prefix_3::suffix_3 ... ok [INFO] [stdout] test should_tokenize_repeat::s_2::prefix_1::suffix_1 ... ok [INFO] [stdout] test should_tokenize_repeat::s_1::prefix_3::suffix_2 ... ok [INFO] [stdout] test should_tokenize_repeat::s_2::prefix_1::suffix_3 ... ok [INFO] [stdout] test should_tokenize_repeat::s_2::prefix_2::suffix_2 ... ok [INFO] [stdout] test should_tokenize_repeat::s_2::prefix_1::suffix_2 ... ok [INFO] [stdout] test should_tokenize_repeat::s_2::prefix_2::suffix_3 ... ok [INFO] [stdout] test should_tokenize_repeat::s_2::prefix_3::suffix_2 ... ok [INFO] [stdout] test should_tokenize_repeat::s_2::prefix_3::suffix_3 ... ok [INFO] [stdout] test should_tokenize_repeat::s_3::prefix_1::suffix_1 ... ok [INFO] [stdout] test should_tokenize_repeat::s_3::prefix_1::suffix_3 ... ok [INFO] [stdout] test should_tokenize_repeat::s_3::prefix_2::suffix_1 ... ok [INFO] [stdout] test should_tokenize_repeat::s_2::prefix_2::suffix_1 ... ok [INFO] [stdout] test should_tokenize_repeat::s_3::prefix_2::suffix_2 ... ok [INFO] [stdout] test should_tokenize_repeat::s_3::prefix_1::suffix_2 ... ok [INFO] [stdout] test should_tokenize_repeat::s_3::prefix_2::suffix_3 ... ok [INFO] [stdout] test should_tokenize_repeat::s_2::prefix_3::suffix_1 ... ok [INFO] [stdout] test should_tokenize_repeat::s_3::prefix_3::suffix_1 ... ok [INFO] [stdout] test should_tokenize_repeat::s_3::prefix_3::suffix_2 ... ok [INFO] [stdout] test should_tokenize_repeat::s_3::prefix_3::suffix_3 ... ok [INFO] [stdout] test should_tokenize_repeat::s_4::prefix_1::suffix_1 ... ok [INFO] [stdout] test should_tokenize_repeat::s_4::prefix_1::suffix_2 ... ok [INFO] [stdout] test should_tokenize_repeat::s_4::prefix_1::suffix_3 ... ok [INFO] [stdout] test should_tokenize_repeat::s_4::prefix_2::suffix_1 ... ok [INFO] [stdout] test should_tokenize_repeat::s_4::prefix_2::suffix_2 ... ok [INFO] [stdout] test should_tokenize_repeat::s_4::prefix_2::suffix_3 ... ok [INFO] [stdout] test should_tokenize_repeat::s_4::prefix_3::suffix_1 ... ok [INFO] [stdout] test should_tokenize_repeat::s_4::prefix_3::suffix_2 ... ok [INFO] [stdout] test should_tokenize_repeat::s_4::prefix_3::suffix_3 ... ok [INFO] [stdout] test should_tokenize_repeat_with_length::len_1 ... ok [INFO] [stdout] test should_tokenize_repeat_with_length::len_2 ... ok [INFO] [stdout] test should_tokenize_repeat_with_length::len_3 ... ok [INFO] [stdout] test should_tokenize_repeat_with_length::len_4 ... ok [INFO] [stdout] test should_treat_empty_equivalence_as_literal ... ok [INFO] [stdout] test single_literal_should_be_of_type_literal::s_1 ... ok [INFO] [stdout] test single_literal_should_be_of_type_literal::s_3 ... ok [INFO] [stdout] test single_literal_should_end_token_stream::s_1 ... ok [INFO] [stdout] test single_literal_should_end_token_stream::s_3 ... ok [INFO] [stdout] test single_literal_should_end_token_stream::s_2 ... ok [INFO] [stdout] test single_literal_should_end_token_stream::s_4 ... ok [INFO] [stdout] test solo_range_should_tokenize_togther ... ok [INFO] [stdout] test single_literal_should_be_of_type_literal::s_2 ... ok [INFO] [stdout] test string_literal_should_be_of_type_literal::s_2 ... ok [INFO] [stdout] test single_literal_should_be_of_type_literal::s_4 ... ok [INFO] [stdout] test solo_range_should_have_type_range ... ok [INFO] [stdout] test string_literal_should_be_of_type_literal::s_1 ... ok [INFO] [stdout] test string_literal_should_be_of_type_literal::s_3 ... ok [INFO] [stdout] test string_literal_should_be_of_type_literal::s_4 ... ok [INFO] [stdout] test string_literal_should_end_token_stream::s_1 ... ok [INFO] [stdout] test string_literal_should_end_token_stream::s_2 ... ok [INFO] [stdout] test string_literal_should_end_token_stream::s_3 ... ok [INFO] [stdout] test string_literal_should_end_token_stream::s_4 ... ok [INFO] [stdout] test tokenizer_should_be_iterator ... ok [INFO] [stdout] test tokenizer_should_return_initial_backslash_escape::s_1 ... ok [INFO] [stdout] test tokenizer_should_return_initial_backslash_escape::s_2 ... ok [INFO] [stdout] test tokenizer_should_return_initial_backslash_escape::s_3 ... ok [INFO] [stdout] test tokenizer_should_return_initial_backslash_escape::s_4 ... ok [INFO] [stdout] test tokenizer_should_return_initial_backslash_escape::s_5 ... ok [INFO] [stdout] test tokenizer_should_return_initial_backslash_escape::s_6 ... ok [INFO] [stdout] test tokenizer_should_return_initial_backslash_escape::s_7 ... ok [INFO] [stdout] test tokenizer_should_return_initial_backslash_escape::s_8 ... ok [INFO] [stdout] test tokenizer_should_return_single_literal::s_1 ... ok [INFO] [stdout] test tokenizer_should_return_single_literal::s_2 ... ok [INFO] [stdout] test tokenizer_should_return_single_literal::s_3 ... ok [INFO] [stdout] test tokenizer_should_return_single_literal::s_4 ... ok [INFO] [stdout] test tokenizer_should_return_string_literal::s_1 ... ok [INFO] [stdout] test tokenizer_should_return_string_literal::s_2 ... ok [INFO] [stdout] test tokenizer_should_return_string_literal::s_3 ... ok [INFO] [stdout] test tokenizer_should_return_string_literal::s_4 ... ok [INFO] [stdout] test tr_actual_treats_repeat_with_cardinality_zero_as_repeat ... ok [INFO] [stdout] test unicode_sequence_should_parse ... ok [INFO] [stdout] [INFO] [stdout] test result: ok. 253 passed; 0 failed; 47 ignored; 0 measured; 0 filtered out; finished in 0.08s [INFO] [stdout] [INFO] [stderr] Running tests/map_charsets.rs (/opt/rustwide/target/x86_64-unknown-linux-musl/debug/deps/map_charsets-7e5cd8bc043a19b1) [INFO] [stdout] [INFO] [stdout] running 22 tests [INFO] [stdout] test should_extend_set2_to_set1_length ... ignored [INFO] [stdout] test mapping_empty_charsets_should_succeed ... ok [INFO] [stdout] test should_map_ascii_to_unicode ... ok [INFO] [stdout] test should_map_escape_in_set2 ... ok [INFO] [stdout] test should_map_escapes_in_set2 ... ok [INFO] [stdout] test should_map_unicode_to_ascii ... ok [INFO] [stdout] test should_map_escapes_in_mixed_set2 ... ok [INFO] [stdout] test should_map_escapes_in_set1 ... ok [INFO] [stdout] test should_map_unicode_to_unicode ... ok [INFO] [stdout] test should_map_unit_length_set1_to_set2 ... ok [INFO] [stdout] test should_map_escapes_in_mixed_set1 ... ok [INFO] [stdout] test should_map_same_length_set1_to_set2 ... ok [INFO] [stdout] test extraneous_chars_in_set2_should_be_ignored ... ok [INFO] [stdout] test should_map_escape_in_set1 ... ok [INFO] [stdout] test unicode_in_mapping_should_preserve_character_length::case_1 ... ok [INFO] [stdout] test unicode_in_mapping_should_preserve_character_length::case_3 ... ok [INFO] [stdout] test unicode_in_mapping_should_preserve_character_length::case_4 ... ok [INFO] [stdout] test unicode_in_mapping_should_preserve_character_length::case_5 ... ok [INFO] [stdout] test unicode_in_mapping_should_preserve_character_length::case_6 ... ok [INFO] [stdout] test unicode_in_mapping_should_preserve_character_length::case_7 ... ok [INFO] [stdout] test unicode_in_mapping_should_preserve_character_length::case_8 ... ok [INFO] [stdout] test unicode_in_mapping_should_preserve_character_length::case_2 ... ok [INFO] [stdout] [INFO] [stdout] test result: ok. 21 passed; 0 failed; 1 ignored; 0 measured; 0 filtered out; finished in 0.01s [INFO] [stdout] [INFO] [stderr] Running tests/parser.rs (/opt/rustwide/target/x86_64-unknown-linux-musl/debug/deps/parser-276b0ea84905056c) [INFO] [stdout] [INFO] [stdout] running 29 tests [INFO] [stdout] test alnum_class_should_expand ... ok [INFO] [stdout] test backslash_9_should_produce_9 ... ok [INFO] [stdout] test backspace_escape_should_produce_backspace ... ok [INFO] [stdout] test backslash_should_produce_backslash ... ok [INFO] [stdout] test bel_escape_should_produce_bel ... ok [INFO] [stdout] test blank_class_should_expand ... ok [INFO] [stdout] test character_range_should_produce_characters::case_1 ... ok [INFO] [stdout] test character_range_should_produce_characters::case_2 ... ok [INFO] [stdout] test character_range_should_produce_characters::case_3 ... ok [INFO] [stdout] test character_range_should_produce_characters::case_4 ... ok [INFO] [stdout] test character_range_should_produce_characters::case_5 ... ok [INFO] [stdout] test character_range_should_produce_characters::case_6 ... ok [INFO] [stdout] test character_range_used_as_verbose_spelling_for_character ... ok [INFO] [stdout] test cntrl_class_should_expand ... ok [INFO] [stdout] test backslash_8_should_produce_8 ... ok [INFO] [stdout] test alpha_class_should_expand ... ok [INFO] [stdout] test cr_escape_should_produce_cr ... ok [INFO] [stdout] test digit_class_should_expand ... ok [INFO] [stdout] test graph_class_should_expand ... ok [INFO] [stdout] test lower_class_should_expand ... ok [INFO] [stdout] test newline_escape_should_produce_newline ... ok [INFO] [stdout] test print_class_should_expand ... ok [INFO] [stdout] test space_class_should_expand ... ok [INFO] [stdout] test tab_escape_should_produce_tab ... ok [INFO] [stdout] test vertical_tab_escape_should_produce_vertical_tab ... ok [INFO] [stdout] test xdigit_class_should_expand ... ok [INFO] [stdout] test formfeed_escape_should_produce_formfeed ... ok [INFO] [stdout] test punct_class_should_expand ... ok [INFO] [stderr] Running tests/rpad.rs (/opt/rustwide/target/x86_64-unknown-linux-musl/debug/deps/rpad-e9c78f61218fb619) [INFO] [stdout] test upper_class_should_expand ... ok [INFO] [stdout] [INFO] [stdout] test result: ok. 29 passed; 0 failed; 0 ignored; 0 measured; 0 filtered out; finished in 0.01s [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] running 6 tests [INFO] [stdout] test should_extend_str ... ok [INFO] [stdout] test should_extend_unit_str_by_one ... ok [INFO] [stdout] test should_extend_unit_str_by_two ... ok [INFO] [stdout] test should_ignore_length_less_than_input_str_length ... ok [INFO] [stdout] test should_return_input_str_with_of_supplied_length ... ok [INFO] [stdout] test should_succeed_with_empty_string ... ok [INFO] [stdout] [INFO] [stdout] test result: ok. 6 passed; 0 failed; 0 ignored; 0 measured; 0 filtered out; finished in 0.00s [INFO] [stdout] [INFO] running `Command { std: "docker" "inspect" "11a066a041df87198b815c2f5ee75436cf200c6010baa11d920d65c48c23fe4b", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "11a066a041df87198b815c2f5ee75436cf200c6010baa11d920d65c48c23fe4b", kill_on_drop: false }` [INFO] [stdout] 11a066a041df87198b815c2f5ee75436cf200c6010baa11d920d65c48c23fe4b