[INFO] updating cached repository johnny-human/nlp-tokenize [INFO] running `"git" "fetch" "--all"` [INFO] [stdout] Fetching origin [INFO] [stderr] From git://github.com/johnny-human/nlp-tokenize [INFO] [stderr] * branch HEAD -> FETCH_HEAD [INFO] running `"git" "clone" "work/cache/sources/gh/johnny-human/nlp-tokenize" "work/ex/pr-58948/sources/master#0f88167f89fffe321590c5148f21b7d51d44388d/gh/johnny-human/nlp-tokenize"` [INFO] [stderr] Cloning into 'work/ex/pr-58948/sources/master#0f88167f89fffe321590c5148f21b7d51d44388d/gh/johnny-human/nlp-tokenize'... [INFO] [stderr] done. [INFO] running `"git" "clone" "work/cache/sources/gh/johnny-human/nlp-tokenize" "work/ex/pr-58948/sources/try#a829e31f924ec11e400e8d0c2fd2c4bb8937ae6b/gh/johnny-human/nlp-tokenize"` [INFO] [stderr] Cloning into 'work/ex/pr-58948/sources/try#a829e31f924ec11e400e8d0c2fd2c4bb8937ae6b/gh/johnny-human/nlp-tokenize'... [INFO] [stderr] done. [INFO] running `"git" "rev-parse" "HEAD"` [INFO] [stdout] fa68c7ce430a43fd9b9f7983f97a58ec4c195caf [INFO] sha for GitHub repo johnny-human/nlp-tokenize: fa68c7ce430a43fd9b9f7983f97a58ec4c195caf [INFO] validating manifest of johnny-human/nlp-tokenize on toolchain master#0f88167f89fffe321590c5148f21b7d51d44388d [INFO] running `"/mnt/big/crater/work/local/cargo-home/bin/cargo" "+0f88167f89fffe321590c5148f21b7d51d44388d-alt" "read-manifest" "--manifest-path" "Cargo.toml"` [INFO] validating manifest of johnny-human/nlp-tokenize on toolchain try#a829e31f924ec11e400e8d0c2fd2c4bb8937ae6b [INFO] running `"/mnt/big/crater/work/local/cargo-home/bin/cargo" "+a829e31f924ec11e400e8d0c2fd2c4bb8937ae6b-alt" "read-manifest" "--manifest-path" "Cargo.toml"` [INFO] started frobbing johnny-human/nlp-tokenize [INFO] finished frobbing johnny-human/nlp-tokenize [INFO] frobbed toml for johnny-human/nlp-tokenize written to work/ex/pr-58948/sources/master#0f88167f89fffe321590c5148f21b7d51d44388d/gh/johnny-human/nlp-tokenize/Cargo.toml [INFO] started frobbing johnny-human/nlp-tokenize [INFO] finished frobbing johnny-human/nlp-tokenize [INFO] frobbed toml for johnny-human/nlp-tokenize written to work/ex/pr-58948/sources/try#a829e31f924ec11e400e8d0c2fd2c4bb8937ae6b/gh/johnny-human/nlp-tokenize/Cargo.toml [INFO] crate johnny-human/nlp-tokenize already has a lockfile, it will not be regenerated [INFO] running `"/mnt/big/crater/work/local/cargo-home/bin/cargo" "+0f88167f89fffe321590c5148f21b7d51d44388d-alt" "fetch" "--locked" "--manifest-path" "Cargo.toml"` [INFO] running `"/mnt/big/crater/work/local/cargo-home/bin/cargo" "+a829e31f924ec11e400e8d0c2fd2c4bb8937ae6b-alt" "fetch" "--locked" "--manifest-path" "Cargo.toml"` [INFO] checking johnny-human/nlp-tokenize against try#a829e31f924ec11e400e8d0c2fd2c4bb8937ae6b for pr-58948 [INFO] running `"docker" "create" "-v" "/mnt/big/crater/work/local/target-dirs/pr-58948/worker-7/try#a829e31f924ec11e400e8d0c2fd2c4bb8937ae6b:/opt/crater/target:rw,Z" "-v" "/mnt/big/crater/work/ex/pr-58948/sources/try#a829e31f924ec11e400e8d0c2fd2c4bb8937ae6b/gh/johnny-human/nlp-tokenize:/opt/crater/workdir:ro,Z" "-v" "/mnt/big/crater/work/local/cargo-home:/opt/crater/cargo-home:ro,Z" "-v" "/mnt/big/crater/work/local/rustup-home:/opt/crater/rustup-home:ro,Z" "-e" "USER_ID=1000" "-e" "SOURCE_DIR=/opt/crater/workdir" "-e" "MAP_USER_ID=1000" "-e" "CARGO_TARGET_DIR=/opt/crater/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/crater/cargo-home" "-e" "RUSTUP_HOME=/opt/crater/rustup-home" "-w" "/opt/crater/workdir" "-m" "1536M" "--network" "none" "rustops/crates-build-env" "/opt/crater/cargo-home/bin/cargo" "+a829e31f924ec11e400e8d0c2fd2c4bb8937ae6b-alt" "check" "--frozen" "--all" "--all-targets"` [INFO] [stdout] 4da560982ad54f8e99ea5baab818fa634fba09fcf282c523b1559b147a6f93ef [INFO] running `"docker" "start" "-a" "4da560982ad54f8e99ea5baab818fa634fba09fcf282c523b1559b147a6f93ef"` [INFO] [stderr] Checking bitflags v0.9.0 [INFO] [stderr] Checking clap v2.24.1 [INFO] [stderr] Checking fs2 v0.2.5 [INFO] [stderr] Checking filebuffer v0.1.1 [INFO] [stderr] Checking memmap v0.4.0 [INFO] [stderr] Checking fst v0.1.38 [INFO] [stderr] Checking nlp v0.1.0 (/opt/crater/workdir) [INFO] [stderr] warning: unused import: `std::borrow::Cow` [INFO] [stderr] --> src/tokenizer_loop.rs:2:5 [INFO] [stderr] | [INFO] [stderr] 2 | use std::borrow::Cow; [INFO] [stderr] | ^^^^^^^^^^^^^^^^ [INFO] [stderr] | [INFO] [stderr] = note: #[warn(unused_imports)] on by default [INFO] [stderr] [INFO] [stderr] warning: unused imports: `DecoderTrap`, `Encoding` [INFO] [stderr] --> src/tokenizer_loop.rs:3:16 [INFO] [stderr] | [INFO] [stderr] 3 | use encoding::{Encoding, DecoderTrap}; [INFO] [stderr] | ^^^^^^^^ ^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused import: `encoding::all::UTF_8` [INFO] [stderr] --> src/tokenizer_loop.rs:4:5 [INFO] [stderr] | [INFO] [stderr] 4 | use encoding::all::UTF_8; [INFO] [stderr] | ^^^^^^^^^^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused import: `bitflags::*` [INFO] [stderr] --> src/numbers.rs:1:5 [INFO] [stderr] | [INFO] [stderr] 1 | use bitflags::*; [INFO] [stderr] | ^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused import: `std::borrow::Cow` [INFO] [stderr] --> src/numbers.rs:2:5 [INFO] [stderr] | [INFO] [stderr] 2 | use std::borrow::Cow; [INFO] [stderr] | ^^^^^^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused imports: `DecoderTrap`, `Encoding` [INFO] [stderr] --> src/numbers.rs:3:16 [INFO] [stderr] | [INFO] [stderr] 3 | use encoding::{Encoding, DecoderTrap}; [INFO] [stderr] | ^^^^^^^^ ^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused import: `encoding::all::UTF_8` [INFO] [stderr] --> src/numbers.rs:4:5 [INFO] [stderr] | [INFO] [stderr] 4 | use encoding::all::UTF_8; [INFO] [stderr] | ^^^^^^^^^^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused import: `Duration` [INFO] [stderr] --> src/main.rs:14:25 [INFO] [stderr] | [INFO] [stderr] 14 | use time::{PreciseTime, Duration}; [INFO] [stderr] | ^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused import: `tokenizer_peek::*` [INFO] [stderr] --> src/main.rs:16:5 [INFO] [stderr] | [INFO] [stderr] 16 | use tokenizer_peek::*; [INFO] [stderr] | ^^^^^^^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused import: `std::borrow::Cow` [INFO] [stderr] --> src/tokenizer_loop.rs:2:5 [INFO] [stderr] | [INFO] [stderr] 2 | use std::borrow::Cow; [INFO] [stderr] | ^^^^^^^^^^^^^^^^ [INFO] [stderr] | [INFO] [stderr] = note: #[warn(unused_imports)] on by default [INFO] [stderr] [INFO] [stderr] warning: unused imports: `DecoderTrap`, `Encoding` [INFO] [stderr] --> src/tokenizer_loop.rs:3:16 [INFO] [stderr] | [INFO] [stderr] 3 | use encoding::{Encoding, DecoderTrap}; [INFO] [stderr] | ^^^^^^^^ ^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused import: `encoding::all::UTF_8` [INFO] [stderr] --> src/tokenizer_loop.rs:4:5 [INFO] [stderr] | [INFO] [stderr] 4 | use encoding::all::UTF_8; [INFO] [stderr] | ^^^^^^^^^^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused import: `bitflags::*` [INFO] [stderr] --> src/numbers.rs:1:5 [INFO] [stderr] | [INFO] [stderr] 1 | use bitflags::*; [INFO] [stderr] | ^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused import: `std::borrow::Cow` [INFO] [stderr] --> src/numbers.rs:2:5 [INFO] [stderr] | [INFO] [stderr] 2 | use std::borrow::Cow; [INFO] [stderr] | ^^^^^^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused imports: `DecoderTrap`, `Encoding` [INFO] [stderr] --> src/numbers.rs:3:16 [INFO] [stderr] | [INFO] [stderr] 3 | use encoding::{Encoding, DecoderTrap}; [INFO] [stderr] | ^^^^^^^^ ^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused import: `encoding::all::UTF_8` [INFO] [stderr] --> src/numbers.rs:4:5 [INFO] [stderr] | [INFO] [stderr] 4 | use encoding::all::UTF_8; [INFO] [stderr] | ^^^^^^^^^^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused import: `Duration` [INFO] [stderr] --> src/main.rs:14:25 [INFO] [stderr] | [INFO] [stderr] 14 | use time::{PreciseTime, Duration}; [INFO] [stderr] | ^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused import: `tokenizer_peek::*` [INFO] [stderr] --> src/main.rs:16:5 [INFO] [stderr] | [INFO] [stderr] 16 | use tokenizer_peek::*; [INFO] [stderr] | ^^^^^^^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused variable: `token` [INFO] [stderr] --> src/tokenizer_loop.rs:97:17 [INFO] [stderr] | [INFO] [stderr] 97 | let token = Token { [INFO] [stderr] | ^^^^^ help: consider prefixing with an underscore: `_token` [INFO] [stderr] | [INFO] [stderr] = note: #[warn(unused_variables)] on by default [INFO] [stderr] [INFO] [stderr] warning: unused variable: `length` [INFO] [stderr] --> src/tokenizer_peek.rs:75:11 [INFO] [stderr] | [INFO] [stderr] 75 | let length = c.len(); [INFO] [stderr] | ^^^^^^ help: consider prefixing with an underscore: `_length` [INFO] [stderr] [INFO] [stderr] warning: unused variable: `i` [INFO] [stderr] --> src/tokenizer_peek.rs:77:12 [INFO] [stderr] | [INFO] [stderr] 77 | for (i, byte) in c.iter().enumerate() { [INFO] [stderr] | ^ help: consider prefixing with an underscore: `_i` [INFO] [stderr] [INFO] [stderr] warning: value assigned to `pos` is never read [INFO] [stderr] --> src/numbers.rs:11:15 [INFO] [stderr] | [INFO] [stderr] 11 | let mut pos: usize = 0; [INFO] [stderr] | ^^^ [INFO] [stderr] | [INFO] [stderr] = note: #[warn(unused_assignments)] on by default [INFO] [stderr] = help: maybe it is overwritten before being read? [INFO] [stderr] [INFO] [stderr] warning: unused variable: `token` [INFO] [stderr] --> src/tokenizer_loop.rs:97:17 [INFO] [stderr] | [INFO] [stderr] 97 | let token = Token { [INFO] [stderr] | ^^^^^ help: consider prefixing with an underscore: `_token` [INFO] [stderr] | [INFO] [stderr] = note: #[warn(unused_variables)] on by default [INFO] [stderr] [INFO] [stderr] warning: unused variable: `length` [INFO] [stderr] --> src/tokenizer_peek.rs:75:11 [INFO] [stderr] | [INFO] [stderr] 75 | let length = c.len(); [INFO] [stderr] | ^^^^^^ help: consider prefixing with an underscore: `_length` [INFO] [stderr] [INFO] [stderr] warning: unused variable: `i` [INFO] [stderr] --> src/tokenizer_peek.rs:77:12 [INFO] [stderr] | [INFO] [stderr] 77 | for (i, byte) in c.iter().enumerate() { [INFO] [stderr] | ^ help: consider prefixing with an underscore: `_i` [INFO] [stderr] [INFO] [stderr] warning: value assigned to `pos` is never read [INFO] [stderr] --> src/numbers.rs:11:15 [INFO] [stderr] | [INFO] [stderr] 11 | let mut pos: usize = 0; [INFO] [stderr] | ^^^ [INFO] [stderr] | [INFO] [stderr] = note: #[warn(unused_assignments)] on by default [INFO] [stderr] = help: maybe it is overwritten before being read? [INFO] [stderr] [INFO] [stderr] warning: variant is never constructed: `Alpha` [INFO] [stderr] --> src/tokenizer_peek.rs:16:7 [INFO] [stderr] | [INFO] [stderr] 16 | Alpha, // Pure alpha string [INFO] [stderr] | ^^^^^ [INFO] [stderr] | [INFO] [stderr] = note: #[warn(dead_code)] on by default [INFO] [stderr] [INFO] [stderr] warning: variant is never constructed: `Other` [INFO] [stderr] --> src/tokenizer_peek.rs:17:7 [INFO] [stderr] | [INFO] [stderr] 17 | Other, // Unidentified elements [INFO] [stderr] | ^^^^^ [INFO] [stderr] [INFO] [stderr] warning: variant is never constructed: `Numeric` [INFO] [stderr] --> src/tokenizer_peek.rs:21:7 [INFO] [stderr] | [INFO] [stderr] 21 | Numeric, // Pure numeric strings 123 [INFO] [stderr] | ^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: variant is never constructed: `NonAlphaNum` [INFO] [stderr] --> src/tokenizer_peek.rs:22:7 [INFO] [stderr] | [INFO] [stderr] 22 | NonAlphaNum // Strings with only | ¦ § _ ~ ^ [INFO] [stderr] | ^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: function `U80_FF` should have a snake case name [INFO] [stderr] --> src/tokenizer_loop.rs:141:8 [INFO] [stderr] | [INFO] [stderr] 141 | pub fn U80_FF(byte: u8) -> [u8; 2] { [INFO] [stderr] | ^^^^^^ help: convert the identifier to snake case: `u80_ff` [INFO] [stderr] | [INFO] [stderr] = note: #[warn(non_snake_case)] on by default [INFO] [stderr] [INFO] [stderr] warning: variant is never constructed: `Alpha` [INFO] [stderr] --> src/tokenizer_peek.rs:16:7 [INFO] [stderr] | [INFO] [stderr] 16 | Alpha, // Pure alpha string [INFO] [stderr] | ^^^^^ [INFO] [stderr] | [INFO] [stderr] = note: #[warn(dead_code)] on by default [INFO] [stderr] [INFO] [stderr] warning: variant is never constructed: `Other` [INFO] [stderr] --> src/tokenizer_peek.rs:17:7 [INFO] [stderr] | [INFO] [stderr] 17 | Other, // Unidentified elements [INFO] [stderr] | ^^^^^ [INFO] [stderr] [INFO] [stderr] warning: variant is never constructed: `Numeric` [INFO] [stderr] --> src/tokenizer_peek.rs:21:7 [INFO] [stderr] | [INFO] [stderr] 21 | Numeric, // Pure numeric strings 123 [INFO] [stderr] | ^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: variant is never constructed: `NonAlphaNum` [INFO] [stderr] --> src/tokenizer_peek.rs:22:7 [INFO] [stderr] | [INFO] [stderr] 22 | NonAlphaNum // Strings with only | ¦ § _ ~ ^ [INFO] [stderr] | ^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: function `U80_FF` should have a snake case name [INFO] [stderr] --> src/tokenizer_loop.rs:141:8 [INFO] [stderr] | [INFO] [stderr] 141 | pub fn U80_FF(byte: u8) -> [u8; 2] { [INFO] [stderr] | ^^^^^^ help: convert the identifier to snake case: `u80_ff` [INFO] [stderr] | [INFO] [stderr] = note: #[warn(non_snake_case)] on by default [INFO] [stderr] [INFO] [stderr] Finished dev [unoptimized + debuginfo] target(s) in 6.18s [INFO] running `"docker" "inspect" "4da560982ad54f8e99ea5baab818fa634fba09fcf282c523b1559b147a6f93ef"` [INFO] running `"docker" "rm" "-f" "4da560982ad54f8e99ea5baab818fa634fba09fcf282c523b1559b147a6f93ef"` [INFO] [stdout] 4da560982ad54f8e99ea5baab818fa634fba09fcf282c523b1559b147a6f93ef