[INFO] fetching crate rten-text 0.23.0...
[INFO] testing rten-text-0.23.0 against master#c90bcb9571b7aab0d8beaa2ce8a998ffaf079d38 for pr-146098-8
[INFO] extracting crate rten-text 0.23.0 into /workspace/builds/worker-2-tc1/source
[INFO] started tweaking crates.io crate rten-text 0.23.0
[INFO] finished tweaking crates.io crate rten-text 0.23.0
[INFO] tweaked toml for crates.io crate rten-text 0.23.0 written to /workspace/builds/worker-2-tc1/source/Cargo.toml
[INFO] validating manifest of crates.io crate rten-text 0.23.0 on toolchain c90bcb9571b7aab0d8beaa2ce8a998ffaf079d38
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+c90bcb9571b7aab0d8beaa2ce8a998ffaf079d38" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }`
[INFO] crate crates.io crate rten-text 0.23.0 already has a lockfile, it will not be regenerated
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+c90bcb9571b7aab0d8beaa2ce8a998ffaf079d38" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }`
[INFO] [stderr]  Downloading crates ...
[INFO] [stderr]   Downloaded fancy-regex v0.14.0
[INFO] [stderr]   Downloaded syn v2.0.72
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-2-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-2-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:4848fb76d95f26979359cc7e45710b1dbc8f3acb7aeedee7c460d7702230f228" "/opt/rustwide/cargo-home/bin/cargo" "+c90bcb9571b7aab0d8beaa2ce8a998ffaf079d38" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }`
[INFO] [stdout] 43f3341fb9e1f50e57b13d91810bb9a28cbeb69d00f6c52a0e28feb970199fe1
[INFO] running `Command { std: "docker" "start" "-a" "43f3341fb9e1f50e57b13d91810bb9a28cbeb69d00f6c52a0e28feb970199fe1", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "inspect" "43f3341fb9e1f50e57b13d91810bb9a28cbeb69d00f6c52a0e28feb970199fe1", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "43f3341fb9e1f50e57b13d91810bb9a28cbeb69d00f6c52a0e28feb970199fe1", kill_on_drop: false }`
[INFO] [stdout] 43f3341fb9e1f50e57b13d91810bb9a28cbeb69d00f6c52a0e28feb970199fe1
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-2-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-2-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:4848fb76d95f26979359cc7e45710b1dbc8f3acb7aeedee7c460d7702230f228" "/opt/rustwide/cargo-home/bin/cargo" "+c90bcb9571b7aab0d8beaa2ce8a998ffaf079d38" "build" "--frozen" "--message-format=json", kill_on_drop: false }`
[INFO] [stdout] 2363a3392ffb2db60ca8fad4c0fd7128fb269085a379ae443ef5a8745d765587
[INFO] running `Command { std: "docker" "start" "-a" "2363a3392ffb2db60ca8fad4c0fd7128fb269085a379ae443ef5a8745d765587", kill_on_drop: false }`
[INFO] [stderr]    Compiling proc-macro2 v1.0.86
[INFO] [stderr]    Compiling serde v1.0.204
[INFO] [stderr]    Compiling bit-vec v0.8.0
[INFO] [stderr]    Compiling serde_json v1.0.120
[INFO] [stderr]    Compiling regex-syntax v0.8.4
[INFO] [stderr]    Compiling ryu v1.0.18
[INFO] [stderr]    Compiling tinyvec v1.8.0
[INFO] [stderr]    Compiling bit-set v0.8.0
[INFO] [stderr]    Compiling quote v1.0.36
[INFO] [stderr]    Compiling unicode-normalization v0.1.23
[INFO] [stderr]    Compiling syn v2.0.72
[INFO] [stderr]    Compiling regex-automata v0.4.7
[INFO] [stderr]    Compiling serde_derive v1.0.204
[INFO] [stderr]    Compiling fancy-regex v0.14.0
[INFO] [stderr]    Compiling rten-text v0.23.0 (/opt/rustwide/workdir)
[INFO] [stderr]     Finished `dev` profile [unoptimized + debuginfo] target(s) in 14.16s
[INFO] running `Command { std: "docker" "inspect" "2363a3392ffb2db60ca8fad4c0fd7128fb269085a379ae443ef5a8745d765587", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "2363a3392ffb2db60ca8fad4c0fd7128fb269085a379ae443ef5a8745d765587", kill_on_drop: false }`
[INFO] [stdout] 2363a3392ffb2db60ca8fad4c0fd7128fb269085a379ae443ef5a8745d765587
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-2-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-2-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:4848fb76d95f26979359cc7e45710b1dbc8f3acb7aeedee7c460d7702230f228" "/opt/rustwide/cargo-home/bin/cargo" "+c90bcb9571b7aab0d8beaa2ce8a998ffaf079d38" "test" "--frozen" "--no-run" "--message-format=json", kill_on_drop: false }`
[INFO] [stdout] f228ae55699349c7ecdab8c47ada33ddf35677584a01dbf1a5c1516ecd667b0f
[INFO] running `Command { std: "docker" "start" "-a" "f228ae55699349c7ecdab8c47ada33ddf35677584a01dbf1a5c1516ecd667b0f", kill_on_drop: false }`
[INFO] [stderr]    Compiling rten-text v0.23.0 (/opt/rustwide/workdir)
[INFO] [stdout] error[E0432]: unresolved import `rten_testing`
[INFO] [stdout]    --> src/models/bpe.rs:541:9
[INFO] [stdout]     |
[INFO] [stdout] 541 |     use rten_testing::TestCases;
[INFO] [stdout]     |         ^^^^^^^^^^^^ use of unresolved module or unlinked crate `rten_testing`
[INFO] [stdout]     |
[INFO] [stdout]     = help: if you wanted to use a crate named `rten_testing`, use `cargo add rten_testing` to add it to your `Cargo.toml`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0432]: unresolved import `rten_testing`
[INFO] [stdout]    --> src/models/wordpiece.rs:135:9
[INFO] [stdout]     |
[INFO] [stdout] 135 |     use rten_testing::TestCases;
[INFO] [stdout]     |         ^^^^^^^^^^^^ use of unresolved module or unlinked crate `rten_testing`
[INFO] [stdout]     |
[INFO] [stdout]     = help: if you wanted to use a crate named `rten_testing`, use `cargo add rten_testing` to add it to your `Cargo.toml`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0432]: unresolved import `rten_testing`
[INFO] [stdout]    --> src/normalizers.rs:359:9
[INFO] [stdout]     |
[INFO] [stdout] 359 |     use rten_testing::TestCases;
[INFO] [stdout]     |         ^^^^^^^^^^^^ use of unresolved module or unlinked crate `rten_testing`
[INFO] [stdout]     |
[INFO] [stdout]     = help: if you wanted to use a crate named `rten_testing`, use `cargo add rten_testing` to add it to your `Cargo.toml`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0432]: unresolved import `rten_testing`
[INFO] [stdout]    --> src/pre_tokenizers.rs:267:9
[INFO] [stdout]     |
[INFO] [stdout] 267 |     use rten_testing::TestCases;
[INFO] [stdout]     |         ^^^^^^^^^^^^ use of unresolved module or unlinked crate `rten_testing`
[INFO] [stdout]     |
[INFO] [stdout]     = help: if you wanted to use a crate named `rten_testing`, use `cargo add rten_testing` to add it to your `Cargo.toml`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0432]: unresolved import `rten_testing`
[INFO] [stdout]    --> src/tokenizer.rs:835:9
[INFO] [stdout]     |
[INFO] [stdout] 835 |     use rten_testing::TestCases;
[INFO] [stdout]     |         ^^^^^^^^^^^^ use of unresolved module or unlinked crate `rten_testing`
[INFO] [stdout]     |
[INFO] [stdout]     = help: if you wanted to use a crate named `rten_testing`, use `cargo add rten_testing` to add it to your `Cargo.toml`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0432]: unresolved import `rten_testing`
[INFO] [stdout]    --> src/split.rs:156:9
[INFO] [stdout]     |
[INFO] [stdout] 156 |     use rten_testing::TestCases;
[INFO] [stdout]     |         ^^^^^^^^^^^^ use of unresolved module or unlinked crate `rten_testing`
[INFO] [stdout]     |
[INFO] [stdout]     = help: if you wanted to use a crate named `rten_testing`, use `cargo add rten_testing` to add it to your `Cargo.toml`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `test_each` found for array `[bpe::tests::test_encode::Case<'_>; 4]` in the current scope
[INFO] [stdout]    --> src/models/bpe.rs:668:15
[INFO] [stdout]     |
[INFO] [stdout] 668 |         cases.test_each(|case| {
[INFO] [stdout]     |         ------^^^^^^^^^ method not found in `[bpe::tests::test_encode::Case<'_>; 4]`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0277]: the trait bound `str: Into<EncoderInput<'_>>` is not satisfied
[INFO] [stdout]    --> src/models/bpe.rs:688:44
[INFO] [stdout]     |
[INFO] [stdout] 688 |             let encoded = tokenizer.encode(*text, None).unwrap();
[INFO] [stdout]     |                                     ------ ^^^^^ the trait `Into<EncoderInput<'_>>` is not implemented for `str`
[INFO] [stdout]     |                                     |
[INFO] [stdout]     |                                     required by a bound introduced by this call
[INFO] [stdout]     |
[INFO] [stdout]     = note: required for `str` to implement `Into<EncoderInput<'_>>`
[INFO] [stdout] note: required by a bound in `tokenizer::Tokenizer::encode`
[INFO] [stdout]    --> src/tokenizer.rs:500:26
[INFO] [stdout]     |
[INFO] [stdout] 500 |     pub fn encode<'a, I: Into<EncoderInput<'a>>>(
[INFO] [stdout]     |                          ^^^^^^^^^^^^^^^^^^^^^^ required by this bound in `Tokenizer::encode`
[INFO] [stdout] help: consider borrowing here
[INFO] [stdout]     |
[INFO] [stdout] 688 |             let encoded = tokenizer.encode(&*text, None).unwrap();
[INFO] [stdout]     |                                            +
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0277]: the size for values of type `str` cannot be known at compilation time
[INFO] [stdout]    --> src/models/bpe.rs:688:44
[INFO] [stdout]     |
[INFO] [stdout] 688 |             let encoded = tokenizer.encode(*text, None).unwrap();
[INFO] [stdout]     |                                     ------ ^^^^^ doesn't have a size known at compile-time
[INFO] [stdout]     |                                     |
[INFO] [stdout]     |                                     required by a bound introduced by this call
[INFO] [stdout]     |
[INFO] [stdout]     = help: the trait `Sized` is not implemented for `str`
[INFO] [stdout] note: required by an implicit `Sized` bound in `tokenizer::Tokenizer::encode`
[INFO] [stdout]    --> src/tokenizer.rs:500:23
[INFO] [stdout]     |
[INFO] [stdout] 500 |     pub fn encode<'a, I: Into<EncoderInput<'a>>>(
[INFO] [stdout]     |                       ^ required by the implicit `Sized` requirement on this type parameter in `Tokenizer::encode`
[INFO] [stdout] help: references are always `Sized`, even if they point to unsized data; consider not dereferencing the expression
[INFO] [stdout]     |
[INFO] [stdout] 688 -             let encoded = tokenizer.encode(*text, None).unwrap();
[INFO] [stdout] 688 +             let encoded = tokenizer.encode(text, None).unwrap();
[INFO] [stdout]     |
[INFO] [stdout] help: consider relaxing the implicit `Sized` restriction
[INFO] [stdout]    --> src/tokenizer.rs:500:48
[INFO] [stdout]     |
[INFO] [stdout] 500 |     pub fn encode<'a, I: Into<EncoderInput<'a>> + ?Sized>(
[INFO] [stdout]     |                                                 ++++++++
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `test_each` found for array `[bpe::tests::test_get_token_str::Case<'_>; 3]` in the current scope
[INFO] [stdout]    --> src/models/bpe.rs:726:15
[INFO] [stdout]     |
[INFO] [stdout] 726 |         cases.test_each(|case| {
[INFO] [stdout]     |         ------^^^^^^^^^ method not found in `[bpe::tests::test_get_token_str::Case<'_>; 3]`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `test_each` found for array `[bpe::tests::test_decode::Case<'_>; 4]` in the current scope
[INFO] [stdout]    --> src/models/bpe.rs:781:15
[INFO] [stdout]     |
[INFO] [stdout] 781 |         cases.test_each(|case| {
[INFO] [stdout]     |         ------^^^^^^^^^ method not found in `[bpe::tests::test_decode::Case<'_>; 4]`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0277]: the trait bound `str: Into<EncoderInput<'_>>` is not satisfied
[INFO] [stdout]    --> src/models/bpe.rs:801:44
[INFO] [stdout]     |
[INFO] [stdout] 801 |             let encoded = tokenizer.encode(*text, None).unwrap();
[INFO] [stdout]     |                                     ------ ^^^^^ the trait `Into<EncoderInput<'_>>` is not implemented for `str`
[INFO] [stdout]     |                                     |
[INFO] [stdout]     |                                     required by a bound introduced by this call
[INFO] [stdout]     |
[INFO] [stdout]     = note: required for `str` to implement `Into<EncoderInput<'_>>`
[INFO] [stdout] note: required by a bound in `tokenizer::Tokenizer::encode`
[INFO] [stdout]    --> src/tokenizer.rs:500:26
[INFO] [stdout]     |
[INFO] [stdout] 500 |     pub fn encode<'a, I: Into<EncoderInput<'a>>>(
[INFO] [stdout]     |                          ^^^^^^^^^^^^^^^^^^^^^^ required by this bound in `Tokenizer::encode`
[INFO] [stdout] help: consider borrowing here
[INFO] [stdout]     |
[INFO] [stdout] 801 |             let encoded = tokenizer.encode(&*text, None).unwrap();
[INFO] [stdout]     |                                            +
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0277]: the size for values of type `str` cannot be known at compilation time
[INFO] [stdout]    --> src/models/bpe.rs:801:44
[INFO] [stdout]     |
[INFO] [stdout] 801 |             let encoded = tokenizer.encode(*text, None).unwrap();
[INFO] [stdout]     |                                     ------ ^^^^^ doesn't have a size known at compile-time
[INFO] [stdout]     |                                     |
[INFO] [stdout]     |                                     required by a bound introduced by this call
[INFO] [stdout]     |
[INFO] [stdout]     = help: the trait `Sized` is not implemented for `str`
[INFO] [stdout] note: required by an implicit `Sized` bound in `tokenizer::Tokenizer::encode`
[INFO] [stdout]    --> src/tokenizer.rs:500:23
[INFO] [stdout]     |
[INFO] [stdout] 500 |     pub fn encode<'a, I: Into<EncoderInput<'a>>>(
[INFO] [stdout]     |                       ^ required by the implicit `Sized` requirement on this type parameter in `Tokenizer::encode`
[INFO] [stdout] help: references are always `Sized`, even if they point to unsized data; consider not dereferencing the expression
[INFO] [stdout]     |
[INFO] [stdout] 801 -             let encoded = tokenizer.encode(*text, None).unwrap();
[INFO] [stdout] 801 +             let encoded = tokenizer.encode(text, None).unwrap();
[INFO] [stdout]     |
[INFO] [stdout] help: consider relaxing the implicit `Sized` restriction
[INFO] [stdout]    --> src/tokenizer.rs:500:48
[INFO] [stdout]     |
[INFO] [stdout] 500 |     pub fn encode<'a, I: Into<EncoderInput<'a>> + ?Sized>(
[INFO] [stdout]     |                                                 ++++++++
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0614]: type `bool` cannot be dereferenced
[INFO] [stdout]    --> src/models/bpe.rs:803:16
[INFO] [stdout]     |
[INFO] [stdout] 803 |             if *add_eos {
[INFO] [stdout]     |                ^^^^^^^^ can't be dereferenced
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `test_each` found for array `[wordpiece::tests::test_wordpiece_model::Case<'_>; 8]` in the current scope
[INFO] [stdout]    --> src/models/wordpiece.rs:226:15
[INFO] [stdout]     |
[INFO] [stdout] 226 |         cases.test_each(|case| {
[INFO] [stdout]     |         ------^^^^^^^^^ method not found in `[wordpiece::tests::test_wordpiece_model::Case<'_>; 8]`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `test_each` found for array `[wordpiece::tests::test_wordpiece_model_lowercase::Case<'_>; 2]` in the current scope
[INFO] [stdout]    --> src/models/wordpiece.rs:283:15
[INFO] [stdout]     |
[INFO] [stdout] 283 |         cases.test_each(|case| {
[INFO] [stdout]     |         ------^^^^^^^^^ method not found in `[wordpiece::tests::test_wordpiece_model_lowercase::Case<'_>; 2]`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `test_each` found for array `[wordpiece::tests::test_decode::Case<'_>; 3]` in the current scope
[INFO] [stdout]    --> src/models/wordpiece.rs:328:15
[INFO] [stdout]     |
[INFO] [stdout] 328 |         cases.test_each(|case| {
[INFO] [stdout]     |         ------^^^^^^^^^ method not found in `[wordpiece::tests::test_decode::Case<'_>; 3]`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `test_each` found for array `[normalizers::tests::test_bert_lowercase::Case<'_>; 2]` in the current scope
[INFO] [stdout]    --> src/normalizers.rs:412:15
[INFO] [stdout]     |
[INFO] [stdout] 412 |         cases.test_each(|case| {
[INFO] [stdout]     |         ------^^^^^^^^^ method not found in `[normalizers::tests::test_bert_lowercase::Case<'_>; 2]`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `test_each` found for array `[normalizers::tests::test_bert_strip_accepts::Case<'_>; 2]` in the current scope
[INFO] [stdout]    --> src/normalizers.rs:456:15
[INFO] [stdout]     |
[INFO] [stdout] 456 |         cases.test_each(|case| {
[INFO] [stdout]     |         ------^^^^^^^^^ method not found in `[normalizers::tests::test_bert_strip_accepts::Case<'_>; 2]`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0614]: type `bool` cannot be dereferenced
[INFO] [stdout]    --> src/normalizers.rs:465:28
[INFO] [stdout]     |
[INFO] [stdout] 465 |                 lowercase: *lowercase,
[INFO] [stdout]     |                            ^^^^^^^^^^ can't be dereferenced
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `test_each` found for array `[normalizers::tests::test_replace::Case<'_>; 3]` in the current scope
[INFO] [stdout]    --> src/normalizers.rs:514:15
[INFO] [stdout]     |
[INFO] [stdout] 514 |         cases.test_each(|case| {
[INFO] [stdout]     |         ------^^^^^^^^^ method not found in `[normalizers::tests::test_replace::Case<'_>; 3]`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `test_each_value` found for array `[normalizers::tests::test_sequence::Case<'_>; 3]` in the current scope
[INFO] [stdout]    --> src/normalizers.rs:598:15
[INFO] [stdout]     |
[INFO] [stdout] 598 |         cases.test_each_value(|case| {
[INFO] [stdout]     |         ------^^^^^^^^^^^^^^^ method not found in `[normalizers::tests::test_sequence::Case<'_>; 3]`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `test_each` found for array `[normalizers::tests::test_unicode::Case<'_>; 9]` in the current scope
[INFO] [stdout]    --> src/normalizers.rs:672:15
[INFO] [stdout]     |
[INFO] [stdout] 672 |         cases.test_each(|case| {
[INFO] [stdout]     |         ------^^^^^^^^^ method not found in `[normalizers::tests::test_unicode::Case<'_>; 9]`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `test_each` found for array `[pre_tokenizers::tests::test_bert::Case<'_>; 1]` in the current scope
[INFO] [stdout]    --> src/pre_tokenizers.rs:286:15
[INFO] [stdout]     |
[INFO] [stdout] 286 |         cases.test_each(|case| {
[INFO] [stdout]     |         ------^^^^^^^^^ method not found in `[pre_tokenizers::tests::test_bert::Case<'_>; 1]`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `test_each` found for array `[pre_tokenizers::tests::test_digits::Case<'_>; 2]` in the current scope
[INFO] [stdout]    --> src/pre_tokenizers.rs:317:15
[INFO] [stdout]     |
[INFO] [stdout] 317 |         cases.test_each(|case| {
[INFO] [stdout]     |         ------^^^^^^^^^ method not found in `[pre_tokenizers::tests::test_digits::Case<'_>; 2]`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `test_each` found for array `[pre_tokenizers::tests::test_split::Case<'_>; 4]` in the current scope
[INFO] [stdout]    --> src/pre_tokenizers.rs:374:15
[INFO] [stdout]     |
[INFO] [stdout] 374 |         cases.test_each(|case| {
[INFO] [stdout]     |         ------^^^^^^^^^ method not found in `[pre_tokenizers::tests::test_split::Case<'_>; 4]`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `test_each` found for array `[tokenizer::tests::test_text_for_token_range::Case<'_>; 10]` in the current scope
[INFO] [stdout]    --> src/tokenizer.rs:968:15
[INFO] [stdout]     |
[INFO] [stdout] 968 |         cases.test_each(|case| {
[INFO] [stdout]     |         ------^^^^^^^^^ method not found in `[tokenizer::tests::test_text_for_token_range::Case<'_>; 10]`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `test_each` found for array `[tokenizer::tests::test_encode_chunks_single_sequence::Case<'_>; 7]` in the current scope
[INFO] [stdout]     --> src/tokenizer.rs:1085:15
[INFO] [stdout]      |
[INFO] [stdout] 1085 |         cases.test_each(|case| {
[INFO] [stdout]      |         ------^^^^^^^^^ method not found in `[tokenizer::tests::test_encode_chunks_single_sequence::Case<'_>; 7]`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0614]: type `bool` cannot be dereferenced
[INFO] [stdout]     --> src/tokenizer.rs:1104:16
[INFO] [stdout]      |
[INFO] [stdout] 1104 |             if *lowercase {
[INFO] [stdout]      |                ^^^^^^^^^^ can't be dereferenced
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0614]: type `std::option::Option<usize>` cannot be dereferenced
[INFO] [stdout]     --> src/tokenizer.rs:1109:32
[INFO] [stdout]      |
[INFO] [stdout] 1109 |                 max_chunk_len: *max_chunk_len,
[INFO] [stdout]      |                                ^^^^^^^^^^^^^^ can't be dereferenced
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0614]: type `usize` cannot be dereferenced
[INFO] [stdout]     --> src/tokenizer.rs:1110:26
[INFO] [stdout]      |
[INFO] [stdout] 1110 |                 overlap: *overlap,
[INFO] [stdout]      |                          ^^^^^^^^ can't be dereferenced
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `test_each` found for array `[tokenizer::tests::test_encode_chunks_sequence_pair::Case<'_>; 6]` in the current scope
[INFO] [stdout]     --> src/tokenizer.rs:1280:15
[INFO] [stdout]      |
[INFO] [stdout] 1280 |         cases.test_each(|case| {
[INFO] [stdout]      |         ------^^^^^^^^^ method not found in `[tokenizer::tests::test_encode_chunks_sequence_pair::Case<'_>; 6]`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0614]: type `bool` cannot be dereferenced
[INFO] [stdout]     --> src/tokenizer.rs:1300:16
[INFO] [stdout]      |
[INFO] [stdout] 1300 |             if *lowercase {
[INFO] [stdout]      |                ^^^^^^^^^^ can't be dereferenced
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0614]: type `std::option::Option<usize>` cannot be dereferenced
[INFO] [stdout]     --> src/tokenizer.rs:1305:32
[INFO] [stdout]      |
[INFO] [stdout] 1305 |                 max_chunk_len: *max_chunk_len,
[INFO] [stdout]      |                                ^^^^^^^^^^^^^^ can't be dereferenced
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0614]: type `usize` cannot be dereferenced
[INFO] [stdout]     --> src/tokenizer.rs:1306:26
[INFO] [stdout]      |
[INFO] [stdout] 1306 |                 overlap: *overlap,
[INFO] [stdout]      |                          ^^^^^^^^ can't be dereferenced
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0277]: the size for values of type `str` cannot be known at compilation time
[INFO] [stdout]     --> src/tokenizer.rs:1310:32
[INFO] [stdout]      |
[INFO] [stdout] 1310 |                 .encode_chunks((*query, *context).into(), options)
[INFO] [stdout]      |                                ^^^^^^^^^^^^^^^^^^ doesn't have a size known at compile-time
[INFO] [stdout]      |
[INFO] [stdout]      = help: within `(str, str)`, the trait `Sized` is not implemented for `str`
[INFO] [stdout]      = note: required because it appears within the type `(str, str)`
[INFO] [stdout]      = note: tuples must have a statically known size to be initialized
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0277]: the trait bound `EncoderInput<'_>: From<&(str, str)>` is not satisfied
[INFO] [stdout]     --> src/tokenizer.rs:1310:51
[INFO] [stdout]      |
[INFO] [stdout] 1310 |                 .encode_chunks((*query, *context).into(), options)
[INFO] [stdout]      |                                                   ^^^^ unsatisfied trait bound
[INFO] [stdout]      |
[INFO] [stdout] help: the trait `From<&(str, str)>` is not implemented for `EncoderInput<'_>`
[INFO] [stdout]     --> src/tokenizer.rs:33:1
[INFO] [stdout]      |
[INFO] [stdout]   33 | pub enum EncoderInput<'a> {
[INFO] [stdout]      | ^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] help: the following other types implement trait `From<T>`
[INFO] [stdout]     --> src/tokenizer.rs:43:1
[INFO] [stdout]      |
[INFO] [stdout]   43 | impl<'a> From<&'a str> for EncoderInput<'a> {
[INFO] [stdout]      | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ `EncoderInput<'_>` implements `From<&str>`
[INFO] [stdout] ...
[INFO] [stdout]   49 | impl<'a> From<&'a String> for EncoderInput<'a> {
[INFO] [stdout]      | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ `EncoderInput<'_>` implements `From<&std::string::String>`
[INFO] [stdout] ...
[INFO] [stdout]   56 | impl<'a> From<(&'a str, &'a str)> for EncoderInput<'a> {
[INFO] [stdout]      | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ `EncoderInput<'_>` implements `From<(&str, &str)>`
[INFO] [stdout]      = note: required for `&(str, str)` to implement `Into<EncoderInput<'_>>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0614]: type `bool` cannot be dereferenced
[INFO] [stdout]     --> src/tokenizer.rs:1329:39
[INFO] [stdout]      |
[INFO] [stdout] 1329 |                         let text = if *lowercase {
[INFO] [stdout]      |                                       ^^^^^^^^^^ can't be dereferenced
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `test_each` found for array `[split::tests::test_chunks_overlap::Case<'_>; 5]` in the current scope
[INFO] [stdout]    --> src/split.rs:208:15
[INFO] [stdout]     |
[INFO] [stdout] 208 |         cases.test_each(|case| {
[INFO] [stdout]     |         ------^^^^^^^^^ method not found in `[split::tests::test_chunks_overlap::Case<'_>; 5]`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `test_each` found for array `[split::tests::test_split_keep_delimeters::Case<'_>; 5]` in the current scope
[INFO] [stdout]    --> src/split.rs:263:15
[INFO] [stdout]     |
[INFO] [stdout] 263 |         cases.test_each(|case| {
[INFO] [stdout]     |         ------^^^^^^^^^ method not found in `[split::tests::test_split_keep_delimeters::Case<'_>; 5]`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] Some errors have detailed explanations: E0277, E0432, E0599, E0614.
[INFO] [stdout] 
[INFO] [stdout] For more information about an error, try `rustc --explain E0277`.
[INFO] [stdout] 
[INFO] [stderr] error: could not compile `rten-text` (lib test) due to 40 previous errors
[INFO] running `Command { std: "docker" "inspect" "f228ae55699349c7ecdab8c47ada33ddf35677584a01dbf1a5c1516ecd667b0f", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "f228ae55699349c7ecdab8c47ada33ddf35677584a01dbf1a5c1516ecd667b0f", kill_on_drop: false }`
[INFO] [stdout] f228ae55699349c7ecdab8c47ada33ddf35677584a01dbf1a5c1516ecd667b0f
