[INFO] fetching crate mako 0.3.0... [INFO] testing mako-0.3.0 against 1.90.0 for beta-1.91-3 [INFO] extracting crate mako 0.3.0 into /workspace/builds/worker-6-tc1/source [INFO] started tweaking crates.io crate mako 0.3.0 [INFO] finished tweaking crates.io crate mako 0.3.0 [INFO] tweaked toml for crates.io crate mako 0.3.0 written to /workspace/builds/worker-6-tc1/source/Cargo.toml [INFO] validating manifest of crates.io crate mako 0.3.0 on toolchain 1.90.0 [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+1.90.0" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }` [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+1.90.0" "generate-lockfile" "--manifest-path" "Cargo.toml", kill_on_drop: false }` [INFO] [stderr] Updating crates.io index [INFO] [stderr] Locking 139 packages to latest compatible versions [INFO] [stderr] Adding clap v2.34.0 (available: v4.5.48) [INFO] [stderr] Adding criterion v0.3.6 (available: v0.7.0) [INFO] [stderr] Adding derive_builder v0.9.0 (available: v0.20.2) [INFO] [stderr] Adding indicatif v0.15.0 (available: v0.18.0) [INFO] [stderr] Adding itertools v0.9.0 (available: v0.14.0) [INFO] [stderr] Adding rand v0.8.5 (available: v0.9.2) [INFO] [stderr] Adding rayon-cond v0.2.0 (available: v0.4.0) [INFO] [stderr] Adding regex-syntax v0.6.29 (available: v0.8.6) [INFO] [stderr] Adding wasi v0.10.0+wasi-snapshot-preview1 (available: v0.10.2+wasi-snapshot-preview1) [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+1.90.0" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }` [INFO] [stderr] Downloading crates ... [INFO] [stderr] Downloaded thread-control v0.1.2 [INFO] [stderr] Downloaded onig v6.5.1 [INFO] [stderr] Downloaded assert_approx_eq v1.1.0 [INFO] [stderr] Downloaded darling_macro v0.10.2 [INFO] [stderr] Downloaded rayon-cond v0.2.0 [INFO] [stderr] Downloaded strsim v0.9.3 [INFO] [stderr] Downloaded darling v0.10.2 [INFO] [stderr] Downloaded derive_builder v0.9.0 [INFO] [stderr] Downloaded darling_core v0.10.2 [INFO] [stderr] Downloaded unicode-normalization-alignments v0.1.12 [INFO] [stderr] Downloaded console v0.16.1 [INFO] [stderr] Downloaded thread-id v3.3.0 [INFO] [stderr] Downloaded derive_builder_core v0.9.0 [INFO] [stderr] Downloaded owning_ref v0.2.4 [INFO] [stderr] Downloaded parking_lot_core v0.2.14 [INFO] [stderr] Downloaded multiqueue v0.3.2 [INFO] [stderr] Downloaded parking_lot v0.3.8 [INFO] [stderr] Downloaded lentrait v0.2.3 [INFO] [stderr] Downloaded smallvec v0.3.4 [INFO] [stderr] Downloaded esaxx-rs v0.1.10 [INFO] [stderr] Downloaded spm_precompiled v0.1.4 [INFO] [stderr] Downloaded onig_sys v69.9.1 [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-6-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-6-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:e90291280db7d1fac5b66fc6dad9f9662629e7365a55743daf9bdf73ebc4ea79" "/opt/rustwide/cargo-home/bin/cargo" "+1.90.0" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }` [INFO] [stdout] 6ef0fa4826f6260f8371a2df4c9a3f15f167ddaa0cebba7822b6bee0fee2d53a [INFO] running `Command { std: "docker" "start" "-a" "6ef0fa4826f6260f8371a2df4c9a3f15f167ddaa0cebba7822b6bee0fee2d53a", kill_on_drop: false }` [INFO] running `Command { std: "docker" "inspect" "6ef0fa4826f6260f8371a2df4c9a3f15f167ddaa0cebba7822b6bee0fee2d53a", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "6ef0fa4826f6260f8371a2df4c9a3f15f167ddaa0cebba7822b6bee0fee2d53a", kill_on_drop: false }` [INFO] [stdout] 6ef0fa4826f6260f8371a2df4c9a3f15f167ddaa0cebba7822b6bee0fee2d53a [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-6-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-6-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=warn" "-e" "RUSTDOCFLAGS=--cap-lints=warn" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:e90291280db7d1fac5b66fc6dad9f9662629e7365a55743daf9bdf73ebc4ea79" "/opt/rustwide/cargo-home/bin/cargo" "+1.90.0" "build" "--frozen" "--message-format=json", kill_on_drop: false }` [INFO] [stdout] 636955db4977bf9edecf45aa421f3f8306752258e3ee828ece5f5447b5567f9f [INFO] running `Command { std: "docker" "start" "-a" "636955db4977bf9edecf45aa421f3f8306752258e3ee828ece5f5447b5567f9f", kill_on_drop: false }` [INFO] [stderr] Compiling strsim v0.9.3 [INFO] [stderr] Compiling either v1.15.0 [INFO] [stderr] Compiling onig_sys v69.9.1 [INFO] [stderr] Compiling esaxx-rs v0.1.10 [INFO] [stderr] Compiling thread-id v3.3.0 [INFO] [stderr] Compiling minimal-lexical v0.2.1 [INFO] [stderr] Compiling owning_ref v0.2.4 [INFO] [stderr] Compiling derive_builder v0.9.0 [INFO] [stderr] Compiling syn v2.0.106 [INFO] [stderr] Compiling parking_lot_core v0.2.14 [INFO] [stderr] Compiling textwrap v0.11.0 [INFO] [stderr] Compiling console v0.16.1 [INFO] [stderr] Compiling atty v0.2.14 [INFO] [stderr] Compiling smallvec v0.3.4 [INFO] [stderr] Compiling ansi_term v0.12.1 [INFO] [stderr] Compiling vec_map v0.8.2 [INFO] [stderr] Compiling darling_core v0.10.2 [INFO] [stderr] Compiling number_prefix v0.3.0 [INFO] [stderr] Compiling strsim v0.8.0 [INFO] [stderr] Compiling itertools v0.10.5 [INFO] [stderr] Compiling rayon v1.11.0 [INFO] [stderr] Compiling crossbeam v0.2.12 [INFO] [stderr] Compiling itertools v0.9.0 [INFO] [stderr] Compiling nom v7.1.3 [INFO] [stderr] Compiling unicode-normalization-alignments v0.1.12 [INFO] [stderr] Compiling parking_lot v0.3.8 [INFO] [stderr] Compiling indicatif v0.15.0 [INFO] [stderr] Compiling thread-control v0.1.2 [INFO] [stderr] Compiling multiqueue v0.3.2 [INFO] [stderr] Compiling clap v2.34.0 [INFO] [stderr] Compiling regex-syntax v0.6.29 [INFO] [stderr] Compiling unicode_categories v0.1.1 [INFO] [stderr] Compiling lentrait v0.2.3 [INFO] [stderr] Compiling darling_macro v0.10.2 [INFO] [stderr] Compiling darling v0.10.2 [INFO] [stderr] Compiling derive_builder_core v0.9.0 [INFO] [stderr] Compiling rayon-cond v0.2.0 [INFO] [stderr] Compiling onig v6.5.1 [INFO] [stderr] Compiling serde_derive v1.0.228 [INFO] [stderr] Compiling serde v1.0.228 [INFO] [stderr] Compiling spm_precompiled v0.1.4 [INFO] [stderr] Compiling mako v0.3.0 (/opt/rustwide/workdir) [INFO] [stdout] warning: this attribute can only be applied at the crate level [INFO] [stdout] --> src/tokenization/hf_tokenizers/mod.rs:3:8 [INFO] [stdout] | [INFO] [stdout] 3 | #![doc(html_favicon_url = "https://huggingface.co/favicon.ico")] [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: read for more information [INFO] [stdout] = note: `#[warn(invalid_doc_attributes)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: this attribute can only be applied at the crate level [INFO] [stdout] --> src/tokenization/hf_tokenizers/mod.rs:4:8 [INFO] [stdout] | [INFO] [stdout] 4 | #![doc(html_logo_url = "https://huggingface.co/landing/assets/huggingface_logo.svg")] [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: read for more information [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: casting `&T` to `&mut T` is undefined behavior, even if the reference is unused, consider instead using an `UnsafeCell` [INFO] [stdout] --> src/tokenization/hf_tokenizers/models/bpe/trainer.rs:517:47 [INFO] [stdout] | [INFO] [stdout] 513 | let w = &words[*i] as *const _ as *mut _; [INFO] [stdout] | -------------------------------- casting happened here [INFO] [stdout] ... [INFO] [stdout] 517 | let word: &mut Word = &mut (*w); [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: for more information, visit [INFO] [stdout] = note: `#[warn(invalid_reference_casting)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] Finished `dev` profile [unoptimized + debuginfo] target(s) in 21.50s [INFO] running `Command { std: "docker" "inspect" "636955db4977bf9edecf45aa421f3f8306752258e3ee828ece5f5447b5567f9f", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "636955db4977bf9edecf45aa421f3f8306752258e3ee828ece5f5447b5567f9f", kill_on_drop: false }` [INFO] [stdout] 636955db4977bf9edecf45aa421f3f8306752258e3ee828ece5f5447b5567f9f [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-6-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-6-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=warn" "-e" "RUSTDOCFLAGS=--cap-lints=warn" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:e90291280db7d1fac5b66fc6dad9f9662629e7365a55743daf9bdf73ebc4ea79" "/opt/rustwide/cargo-home/bin/cargo" "+1.90.0" "test" "--frozen" "--no-run" "--message-format=json", kill_on_drop: false }` [INFO] [stdout] 3960222dc6238d6514e332145b0fec903e9278a38d07681ee4340fc4614c3e9d [INFO] running `Command { std: "docker" "start" "-a" "3960222dc6238d6514e332145b0fec903e9278a38d07681ee4340fc4614c3e9d", kill_on_drop: false }` [INFO] [stderr] Compiling getrandom v0.3.3 [INFO] [stderr] Compiling plotters-backend v0.3.7 [INFO] [stderr] Compiling csv-core v0.1.12 [INFO] [stderr] Compiling half v1.8.3 [INFO] [stderr] Compiling cast v0.3.0 [INFO] [stderr] Compiling same-file v1.0.6 [INFO] [stderr] Compiling tinytemplate v1.2.1 [INFO] [stderr] Compiling assert_approx_eq v1.1.0 [INFO] [stdout] warning: this attribute can only be applied at the crate level [INFO] [stdout] --> src/tokenization/hf_tokenizers/mod.rs:3:8 [INFO] [stdout] | [INFO] [stdout] 3 | #![doc(html_favicon_url = "https://huggingface.co/favicon.ico")] [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: read for more information [INFO] [stdout] = note: `#[warn(invalid_doc_attributes)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: this attribute can only be applied at the crate level [INFO] [stdout] --> src/tokenization/hf_tokenizers/mod.rs:4:8 [INFO] [stdout] | [INFO] [stdout] 4 | #![doc(html_logo_url = "https://huggingface.co/landing/assets/huggingface_logo.svg")] [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: read for more information [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: casting `&T` to `&mut T` is undefined behavior, even if the reference is unused, consider instead using an `UnsafeCell` [INFO] [stdout] --> src/tokenization/hf_tokenizers/models/bpe/trainer.rs:517:47 [INFO] [stdout] | [INFO] [stdout] 513 | let w = &words[*i] as *const _ as *mut _; [INFO] [stdout] | -------------------------------- casting happened here [INFO] [stdout] ... [INFO] [stdout] 517 | let word: &mut Word = &mut (*w); [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: for more information, visit [INFO] [stdout] = note: `#[warn(invalid_reference_casting)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] Compiling walkdir v2.5.0 [INFO] [stderr] Compiling criterion-plot v0.4.5 [INFO] [stderr] Compiling serde_cbor v0.11.2 [INFO] [stderr] Compiling plotters-svg v0.3.7 [INFO] [stderr] Compiling csv v1.3.1 [INFO] [stderr] Compiling plotters v0.3.7 [INFO] [stderr] Compiling tempfile v3.23.0 [INFO] [stderr] Compiling criterion v0.3.6 [INFO] [stderr] Compiling mako v0.3.0 (/opt/rustwide/workdir) [INFO] [stdout] warning: this attribute can only be applied at the crate level [INFO] [stdout] --> src/tokenization/hf_tokenizers/mod.rs:3:8 [INFO] [stdout] | [INFO] [stdout] 3 | #![doc(html_favicon_url = "https://huggingface.co/favicon.ico")] [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: read for more information [INFO] [stdout] = note: `#[warn(invalid_doc_attributes)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: this attribute can only be applied at the crate level [INFO] [stdout] --> src/tokenization/hf_tokenizers/mod.rs:4:8 [INFO] [stdout] | [INFO] [stdout] 4 | #![doc(html_logo_url = "https://huggingface.co/landing/assets/huggingface_logo.svg")] [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: read for more information [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: casting `&T` to `&mut T` is undefined behavior, even if the reference is unused, consider instead using an `UnsafeCell` [INFO] [stdout] --> src/tokenization/hf_tokenizers/models/bpe/trainer.rs:517:47 [INFO] [stdout] | [INFO] [stdout] 513 | let w = &words[*i] as *const _ as *mut _; [INFO] [stdout] | -------------------------------- casting happened here [INFO] [stdout] ... [INFO] [stdout] 517 | let word: &mut Word = &mut (*w); [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: for more information, visit [INFO] [stdout] = note: `#[warn(invalid_reference_casting)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] Finished `test` profile [unoptimized + debuginfo] target(s) in 18.14s [INFO] running `Command { std: "docker" "inspect" "3960222dc6238d6514e332145b0fec903e9278a38d07681ee4340fc4614c3e9d", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "3960222dc6238d6514e332145b0fec903e9278a38d07681ee4340fc4614c3e9d", kill_on_drop: false }` [INFO] [stdout] 3960222dc6238d6514e332145b0fec903e9278a38d07681ee4340fc4614c3e9d [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-6-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-6-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=warn" "-e" "RUSTDOCFLAGS=--cap-lints=warn" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:e90291280db7d1fac5b66fc6dad9f9662629e7365a55743daf9bdf73ebc4ea79" "/opt/rustwide/cargo-home/bin/cargo" "+1.90.0" "test" "--frozen", kill_on_drop: false }` [INFO] [stdout] f75bd6890244effb2495b7279d978de5c36085e336894cf324be2b42ff5c22f9 [INFO] running `Command { std: "docker" "start" "-a" "f75bd6890244effb2495b7279d978de5c36085e336894cf324be2b42ff5c22f9", kill_on_drop: false }` [INFO] [stderr] warning: this attribute can only be applied at the crate level [INFO] [stderr] --> src/tokenization/hf_tokenizers/mod.rs:3:8 [INFO] [stderr] | [INFO] [stderr] 3 | #![doc(html_favicon_url = "https://huggingface.co/favicon.ico")] [INFO] [stderr] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stderr] | [INFO] [stderr] = note: read for more information [INFO] [stderr] = note: `#[warn(invalid_doc_attributes)]` on by default [INFO] [stderr] [INFO] [stderr] warning: this attribute can only be applied at the crate level [INFO] [stderr] --> src/tokenization/hf_tokenizers/mod.rs:4:8 [INFO] [stderr] | [INFO] [stderr] 4 | #![doc(html_logo_url = "https://huggingface.co/landing/assets/huggingface_logo.svg")] [INFO] [stderr] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stderr] | [INFO] [stderr] = note: read for more information [INFO] [stderr] [INFO] [stderr] warning: casting `&T` to `&mut T` is undefined behavior, even if the reference is unused, consider instead using an `UnsafeCell` [INFO] [stderr] --> src/tokenization/hf_tokenizers/models/bpe/trainer.rs:517:47 [INFO] [stderr] | [INFO] [stderr] 513 | let w = &words[*i] as *const _ as *mut _; [INFO] [stderr] | -------------------------------- casting happened here [INFO] [stderr] ... [INFO] [stderr] 517 | let word: &mut Word = &mut (*w); [INFO] [stderr] | ^^^^^^^^^ [INFO] [stderr] | [INFO] [stderr] = note: for more information, visit [INFO] [stderr] = note: `#[warn(invalid_reference_casting)]` on by default [INFO] [stderr] [INFO] [stderr] warning: `mako` (lib) generated 3 warnings [INFO] [stderr] warning: `mako` (lib test) generated 3 warnings (3 duplicates) [INFO] [stderr] Finished `test` profile [unoptimized + debuginfo] target(s) in 0.18s [INFO] [stderr] Running unittests src/lib.rs (/opt/rustwide/target/debug/deps/mako-2f399f7877e5f2f3) [INFO] [stdout] [INFO] [stdout] running 148 tests [INFO] [stdout] test batching::tests::pad_batch_test ... ok [INFO] [stdout] test batching::tests::pad_mask_test ... ok [INFO] [stdout] test batching::tests::shuffle_lists_test ... ok [INFO] [stdout] test batching::tests::filter_by_length_test ... ok [INFO] [stdout] test pipeline::tests::test_single_pipeline ... ok [INFO] [stdout] test pipeline::tests::test_pair_pipeline ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::unigram::lattice::tests::insert_test ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::unigram::lattice::tests::test_log_sum_exp ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::unigram::lattice::tests::set_sentence ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::tests::trainer_wrapper_train_model_wrapper ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::unigram::lattice::tests::test_populate ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::unigram::lattice::tests::test_nbest ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::unigram::lattice::tests::test_viterbi2 ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::bpe::model::tests::test_bpe_from_file ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::bpe::word::tests::test_merge ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::unigram::model::tests::test_encode2 ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::unigram::serialization::test::test_serialization ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::unigram::serialization::test::test_serialization_no_unk_id ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::unigram::serialization::test::test_serialization_unk_id_not_zero ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::unigram::trainer::tests::test_initial_alphabet ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::bpe::model::tests::test_ordered_vocab_iter ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::unigram::trainer::tests::test_to_log_prob ... ok [INFO] [stdout] test batching::tests::sort_lists_by_length_test ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::unigram::lattice::tests::test_viterbi ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::bpe::model::tests::test_bpe_from_file_bad_merges ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::unigram::model::tests::test_encode ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::wordlevel::serialization::tests::deserialization_should_fail ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::wordlevel::trainer::tests::test_train ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::wordpiece::tests::test_error_display ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::wordpiece::serialization::tests::serde ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::wordpiece::serialization::tests::deserialization_should_fail ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::bpe::model::tests::test_tokenize_with_and_without_dropout ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::bpe::model::tests::test_unk_get_fused ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::bpe::model::tests::test_unk_not_fused ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::bpe::model::tests::test_bpe_with_continuing_subword_prefix ... ok [INFO] [stdout] test tokenization::hf_tokenizers::normalizers::replace::tests::serialization ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::unigram::model::tests::test_populate_nodes_unk ... ok [INFO] [stdout] test tokenization::hf_tokenizers::normalizers::replace::tests::test_replace_regex ... ok [INFO] [stdout] test tokenization::hf_tokenizers::normalizers::strip::tests::test_strip_accents ... ok [INFO] [stdout] test tokenization::hf_tokenizers::normalizers::replace::tests::test_replace ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::bert::tests::basic ... ok [INFO] [stdout] test tokenization::hf_tokenizers::normalizers::strip::tests::test_vietnamese_bug ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::wordlevel::serialization::tests::serde ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::unigram::model::tests::test_populate_nodes ... ok [INFO] [stdout] test tokenization::hf_tokenizers::normalizers::unicode::tests::test_nfkc ... ok [INFO] [stdout] test tokenization::hf_tokenizers::normalizers::strip::tests::test_strip_accents_multiple ... ok [INFO] [stdout] test tokenization::hf_tokenizers::normalizers::strip::tests::test_thai_bug ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::bert::tests::chinese_chars ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::byte_level::tests::processor_trims_offsets ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::digits::tests::individual_digits ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::byte_level::tests::decode_unknown_characters ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::byte_level::tests::decode_works_on_separated_tokens ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::byte_level::tests::decoding ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::byte_level::tests::handling_of_multiple_whitespaces ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::byte_level::tests::handling_of_newlines ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::byte_level::tests::offsets_when_char_split_up ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::byte_level::tests::add_prefix_space ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::metaspace::tests::basic ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::punctuation::tests::punctuation_basic ... ok [INFO] [stdout] test dataloader::tests::test_dataloader ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::bpe::model::tests::test_bpe_from_file_merge_token_oov ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::metaspace::tests::decode ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::metaspace::tests::multiple_spaces ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::sequence::tests::sequence_basic ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::split::tests::basic ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::unicode_scripts::pre_tokenizer::tests::test_unicode_script ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::unicode_scripts::scripts::tests::test_unicode_script ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::unicode_scripts::pre_tokenizer::tests::spaces_are_included_in_every_script ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::unicode_scripts::pre_tokenizer::tests::basic ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::split::tests::regex_string ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::split::tests::serialization ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::digits::tests::numbers ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::split::tests::invert ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::byte_level::tests::pre_tokenization ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::whitespace::tests::whitespace_split ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::bpe::trainer::tests::test_train ... ok [INFO] [stdout] test tokenization::hf_tokenizers::processors::bert::tests::serde ... ok [INFO] [stdout] test tokenization::hf_tokenizers::processors::roberta::tests::serde ... ok [INFO] [stdout] test tokenization::hf_tokenizers::processors::template::tests::pair_must_use_both_sequences ... ok [INFO] [stdout] test tokenization::hf_tokenizers::processors::template::tests::special_token_serde ... ok [INFO] [stdout] test tokenization::hf_tokenizers::processors::template::tests::template_serde ... ok [INFO] [stdout] test tokenization::hf_tokenizers::processors::template::tests::missing_special_tokens ... ok [INFO] [stdout] test tokenization::hf_tokenizers::processors::template::tests::template_processing_serde ... ok [INFO] [stdout] test tokenization::hf_tokenizers::processors::tests::deserialize_bert_roberta_correctly ... ok [INFO] [stdout] test tokenization::hf_tokenizers::processors::template::tests::piece_serde ... ok [INFO] [stdout] test tokenization::hf_tokenizers::processors::template::tests::piece ... ok [INFO] [stdout] test tokenization::hf_tokenizers::processors::template::tests::template_processing ... ok [INFO] [stdout] test tokenization::hf_tokenizers::processors::template::tests::tokens_serde ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::encoding::tests::mappings ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::encoding::tests::merge_encodings ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::encoding::tests::truncate ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::added_vocabulary::tests::empty_matches ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::encoding::tests::truncate_to_empty ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::normalizer::tests::added_around_edges ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::normalizer::tests::added_characters_alignment ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::normalizer::tests::lstrip ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::normalizer::tests::nfd_adds_new_chars ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::normalizer::tests::range_conversion ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::normalizer::tests::append ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::normalizer::tests::remove_at_beginning ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::normalizer::tests::get_range ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::normalizer::tests::original_range ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::normalizer::tests::mixed_addition_and_removal ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::normalizer::tests::remove_chars ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::added_vocabulary::tests::can_add_tokens ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::normalizer::tests::remove_chars_added_by_nfd ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::normalizer::tests::slice ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::normalizer::tests::split ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::normalizer::tests::strip ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::added_vocabulary::tests::can_add_special_tokens ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::normalizer::tests::removed_around_both_edges ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::normalizer::tests::rstrip ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::normalizer::tests::remove_at_end ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::normalizer::tests::replace ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::normalizer::tests::prepend ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::normalizer::tests::transform_check ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::pattern::tests::functions ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::normalizer::tests::transform_range_multiple_bytes ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::pattern::tests::onig_regex ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::pattern::tests::char ... ok [INFO] [stdout] test tokenization::hf_tokenizers::utils::truncation::tests::truncate_encodings_longest_first ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::pattern::tests::str ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::normalizer::tests::transform_range_single_bytes ... ok [INFO] [stdout] test tokenization::hf_tokenizers::utils::truncation::tests::truncate_encodings_empty ... ok [INFO] [stdout] test tokenization::tests::tokenize_alphabet ... ok [INFO] [stdout] test tokenization::tests::tokenize_spaces ... ok [INFO] [stdout] test tokenization::tests::untokenize_alphabet ... ok [INFO] [stdout] test tokenization::tests::untokenize_spaces ... ok [INFO] [stdout] test tokenization::hf_tokenizers::utils::padding::tests::pad_to_multiple ... ok [INFO] [stdout] test tokenization::hf_tokenizers::utils::parallelism::tests::test_maybe_parallel_iterator ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::added_vocabulary::tests::options_use_cases ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::unigram::trainer::tests::test_special_tokens ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::pattern::tests::regex ... ok [INFO] [stdout] test tokenization::hf_tokenizers::tokenizer::added_vocabulary::tests::can_extract_added_tokens ... ok [INFO] [stdout] test tokenization::hf_tokenizers::pre_tokenizers::whitespace::tests::basic ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::unigram::trainer::tests::test_unigram_chars ... ok [INFO] [stdout] test vocab::tests::tokens_from_indexes_wordpiece ... ok [INFO] [stdout] test tokenization::tests::tokenize_wordpiece ... ok [INFO] [stdout] test vocab::tests::indexes_from_tokens_wordpiece ... ok [INFO] [stdout] test tokenization::hf_tokenizers::models::unigram::trainer::tests::test_unk_token ... ok [INFO] [stdout] test tokenization::tests::untokenize_wordpiece ... ok [INFO] [stdout] test vocab::tests::batch_indexes_from_tokens ... ok [INFO] [stdout] test vocab::tests::batch_tokens_from_indexes ... ok [INFO] [stdout] test vocab::tests::indexes_from_tokens_bpe ... ok [INFO] [stdout] test vocab::tests::tokens_from_indexes_bpe ... ok [INFO] [stdout] test vocab::tests::creating_vocab ... ok [INFO] [stdout] test tokenization::tests::tokenize_bpe ... ok [INFO] [stdout] test tokenization::tests::untokenize_bpe ... ok [INFO] [stdout] [INFO] [stdout] test result: ok. 148 passed; 0 failed; 0 ignored; 0 measured; 0 filtered out; finished in 1.27s [INFO] [stdout] [INFO] [stderr] Doc-tests mako [INFO] [stdout] [INFO] [stdout] running 0 tests [INFO] [stdout] [INFO] [stdout] test result: ok. 0 passed; 0 failed; 0 ignored; 0 measured; 0 filtered out; finished in 0.00s [INFO] [stdout] [INFO] running `Command { std: "docker" "inspect" "f75bd6890244effb2495b7279d978de5c36085e336894cf324be2b42ff5c22f9", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "f75bd6890244effb2495b7279d978de5c36085e336894cf324be2b42ff5c22f9", kill_on_drop: false }` [INFO] [stdout] f75bd6890244effb2495b7279d978de5c36085e336894cf324be2b42ff5c22f9