[INFO] updating cached repository johnny-human/nlp-tokenize [INFO] running `"git" "fetch" "--all"` [INFO] [stdout] Fetching origin [INFO] [stderr] From git://github.com/johnny-human/nlp-tokenize [INFO] [stderr] * branch HEAD -> FETCH_HEAD [INFO] running `"git" "clone" "work/cache/sources/gh/johnny-human/nlp-tokenize" "work/ex/pr-63376/sources/master#60960a260f7b5c695fd0717311d72ce62dd4eb43/gh/johnny-human/nlp-tokenize"` [INFO] [stderr] Cloning into 'work/ex/pr-63376/sources/master#60960a260f7b5c695fd0717311d72ce62dd4eb43/gh/johnny-human/nlp-tokenize'... [INFO] [stderr] done. [INFO] running `"git" "clone" "work/cache/sources/gh/johnny-human/nlp-tokenize" "work/ex/pr-63376/sources/try#266783e4e09e4e9d5307c1c8e695659c58bbcac7/gh/johnny-human/nlp-tokenize"` [INFO] [stderr] Cloning into 'work/ex/pr-63376/sources/try#266783e4e09e4e9d5307c1c8e695659c58bbcac7/gh/johnny-human/nlp-tokenize'... [INFO] [stderr] done. [INFO] running `"git" "rev-parse" "HEAD"` [INFO] [stdout] fa68c7ce430a43fd9b9f7983f97a58ec4c195caf [INFO] sha for GitHub repo johnny-human/nlp-tokenize: fa68c7ce430a43fd9b9f7983f97a58ec4c195caf [INFO] validating manifest of johnny-human/nlp-tokenize on toolchain master#60960a260f7b5c695fd0717311d72ce62dd4eb43 [INFO] running `"/mnt/big/crater/work/local/cargo-home/bin/cargo" "+60960a260f7b5c695fd0717311d72ce62dd4eb43-alt" "read-manifest" "--manifest-path" "Cargo.toml"` [INFO] validating manifest of johnny-human/nlp-tokenize on toolchain try#266783e4e09e4e9d5307c1c8e695659c58bbcac7 [INFO] running `"/mnt/big/crater/work/local/cargo-home/bin/cargo" "+266783e4e09e4e9d5307c1c8e695659c58bbcac7-alt" "read-manifest" "--manifest-path" "Cargo.toml"` [INFO] started frobbing johnny-human/nlp-tokenize [INFO] finished frobbing johnny-human/nlp-tokenize [INFO] frobbed toml for johnny-human/nlp-tokenize written to work/ex/pr-63376/sources/master#60960a260f7b5c695fd0717311d72ce62dd4eb43/gh/johnny-human/nlp-tokenize/Cargo.toml [INFO] started frobbing johnny-human/nlp-tokenize [INFO] finished frobbing johnny-human/nlp-tokenize [INFO] frobbed toml for johnny-human/nlp-tokenize written to work/ex/pr-63376/sources/try#266783e4e09e4e9d5307c1c8e695659c58bbcac7/gh/johnny-human/nlp-tokenize/Cargo.toml [INFO] crate johnny-human/nlp-tokenize already has a lockfile, it will not be regenerated [INFO] running `"/mnt/big/crater/work/local/cargo-home/bin/cargo" "+60960a260f7b5c695fd0717311d72ce62dd4eb43-alt" "fetch" "--locked" "--manifest-path" "Cargo.toml"` [INFO] running `"/mnt/big/crater/work/local/cargo-home/bin/cargo" "+266783e4e09e4e9d5307c1c8e695659c58bbcac7-alt" "fetch" "--locked" "--manifest-path" "Cargo.toml"` [INFO] checking johnny-human/nlp-tokenize against try#266783e4e09e4e9d5307c1c8e695659c58bbcac7 for pr-63376 [INFO] running `"docker" "create" "-v" "/mnt/big/crater/work/local/target-dirs/pr-63376/worker-2/try#266783e4e09e4e9d5307c1c8e695659c58bbcac7:/opt/crater/target:rw,Z" "-v" "/mnt/big/crater/work/ex/pr-63376/sources/try#266783e4e09e4e9d5307c1c8e695659c58bbcac7/gh/johnny-human/nlp-tokenize:/opt/crater/workdir:ro,Z" "-v" "/mnt/big/crater/work/local/cargo-home:/opt/crater/cargo-home:ro,Z" "-v" "/mnt/big/crater/work/local/rustup-home:/opt/crater/rustup-home:ro,Z" "-e" "USER_ID=1000" "-e" "SOURCE_DIR=/opt/crater/workdir" "-e" "MAP_USER_ID=1000" "-e" "CARGO_TARGET_DIR=/opt/crater/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/crater/cargo-home" "-e" "RUSTUP_HOME=/opt/crater/rustup-home" "-w" "/opt/crater/workdir" "-m" "1536M" "--network" "none" "rustops/crates-build-env" "/opt/crater/cargo-home/bin/cargo" "+266783e4e09e4e9d5307c1c8e695659c58bbcac7-alt" "check" "--frozen" "--all" "--all-targets"` [INFO] [stdout] 9f9539ca46731fcdeceef45b7e291b2a471eb4fbb922a18c34734397dec07f5f [INFO] running `"docker" "start" "-a" "9f9539ca46731fcdeceef45b7e291b2a471eb4fbb922a18c34734397dec07f5f"` [INFO] [stderr] Checking bitflags v0.9.0 [INFO] [stderr] Checking filebuffer v0.1.1 [INFO] [stderr] Checking fst v0.1.38 [INFO] [stderr] Checking nlp v0.1.0 (/opt/crater/workdir) [INFO] [stderr] warning: unused import: `std::borrow::Cow` [INFO] [stderr] --> src/tokenizer_loop.rs:2:5 [INFO] [stderr] | [INFO] [stderr] 2 | use std::borrow::Cow; [INFO] [stderr] | ^^^^^^^^^^^^^^^^ [INFO] [stderr] | [INFO] [stderr] = note: `#[warn(unused_imports)]` on by default [INFO] [stderr] [INFO] [stderr] warning: unused imports: `DecoderTrap`, `Encoding` [INFO] [stderr] --> src/tokenizer_loop.rs:3:16 [INFO] [stderr] | [INFO] [stderr] 3 | use encoding::{Encoding, DecoderTrap}; [INFO] [stderr] | ^^^^^^^^ ^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused import: `encoding::all::UTF_8` [INFO] [stderr] --> src/tokenizer_loop.rs:4:5 [INFO] [stderr] | [INFO] [stderr] 4 | use encoding::all::UTF_8; [INFO] [stderr] | ^^^^^^^^^^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: trait objects without an explicit `dyn` are deprecated [INFO] [stderr] --> src/tokenizer_loop.rs:36:34 [INFO] [stderr] | [INFO] [stderr] 36 | pub fn words(bytes: Vec, f: &Fn(&Token) -> i32) -> Result, &'static Vec> { [INFO] [stderr] | ^^^^^^^^^^^^^^^^^ help: use `dyn`: `dyn Fn(&Token) -> i32` [INFO] [stderr] | [INFO] [stderr] = note: `#[warn(bare_trait_objects)]` on by default [INFO] [stderr] [INFO] [stderr] warning: `...` range patterns are deprecated [INFO] [stderr] --> src/tokenizer_loop.rs:69:22 [INFO] [stderr] | [INFO] [stderr] 69 | 65 ... 90 | 97 ... 122 => { [INFO] [stderr] | ^^^ help: use `..=` for an inclusive range [INFO] [stderr] | [INFO] [stderr] = note: `#[warn(ellipsis_inclusive_range_patterns)]` on by default [INFO] [stderr] [INFO] [stderr] warning: `...` range patterns are deprecated [INFO] [stderr] --> src/tokenizer_loop.rs:69:34 [INFO] [stderr] | [INFO] [stderr] 69 | 65 ... 90 | 97 ... 122 => { [INFO] [stderr] | ^^^ help: use `..=` for an inclusive range [INFO] [stderr] [INFO] [stderr] warning: `...` range patterns are deprecated [INFO] [stderr] --> src/tokenizer_loop.rs:73:22 [INFO] [stderr] | [INFO] [stderr] 73 | 33 ... 47 | 58 ... 64 | 91 ... 96 | 123 ... 126 | 160 ... 191 => { [INFO] [stderr] | ^^^ help: use `..=` for an inclusive range [INFO] [stderr] [INFO] [stderr] warning: `...` range patterns are deprecated [INFO] [stderr] --> src/tokenizer_loop.rs:73:34 [INFO] [stderr] | [INFO] [stderr] 73 | 33 ... 47 | 58 ... 64 | 91 ... 96 | 123 ... 126 | 160 ... 191 => { [INFO] [stderr] | ^^^ help: use `..=` for an inclusive range [INFO] [stderr] [INFO] [stderr] warning: `...` range patterns are deprecated [INFO] [stderr] --> src/tokenizer_loop.rs:73:46 [INFO] [stderr] | [INFO] [stderr] 73 | 33 ... 47 | 58 ... 64 | 91 ... 96 | 123 ... 126 | 160 ... 191 => { [INFO] [stderr] | ^^^ help: use `..=` for an inclusive range [INFO] [stderr] [INFO] [stderr] warning: `...` range patterns are deprecated [INFO] [stderr] --> src/tokenizer_loop.rs:73:59 [INFO] [stderr] | [INFO] [stderr] 73 | 33 ... 47 | 58 ... 64 | 91 ... 96 | 123 ... 126 | 160 ... 191 => { [INFO] [stderr] | ^^^ help: use `..=` for an inclusive range [INFO] [stderr] [INFO] [stderr] warning: `...` range patterns are deprecated [INFO] [stderr] --> src/tokenizer_loop.rs:73:73 [INFO] [stderr] | [INFO] [stderr] 73 | 33 ... 47 | 58 ... 64 | 91 ... 96 | 123 ... 126 | 160 ... 191 => { [INFO] [stderr] | ^^^ help: use `..=` for an inclusive range [INFO] [stderr] [INFO] [stderr] warning: `...` range patterns are deprecated [INFO] [stderr] --> src/tokenizer_loop.rs:77:22 [INFO] [stderr] | [INFO] [stderr] 77 | 48 ... 57 => { [INFO] [stderr] | ^^^ help: use `..=` for an inclusive range [INFO] [stderr] [INFO] [stderr] warning: `...` range patterns are deprecated [INFO] [stderr] --> src/tokenizer_loop.rs:81:23 [INFO] [stderr] | [INFO] [stderr] 81 | 127 ... 159 | 1 ... 31 => { [INFO] [stderr] | ^^^ help: use `..=` for an inclusive range [INFO] [stderr] [INFO] [stderr] warning: `...` range patterns are deprecated [INFO] [stderr] --> src/tokenizer_loop.rs:81:35 [INFO] [stderr] | [INFO] [stderr] 81 | 127 ... 159 | 1 ... 31 => { [INFO] [stderr] | ^^^ help: use `..=` for an inclusive range [INFO] [stderr] [INFO] [stderr] warning: `...` range patterns are deprecated [INFO] [stderr] --> src/tokenizer_loop.rs:85:23 [INFO] [stderr] | [INFO] [stderr] 85 | 192 ... 255 => { [INFO] [stderr] | ^^^ help: use `..=` for an inclusive range [INFO] [stderr] [INFO] [stderr] warning: unused import: `bitflags::*` [INFO] [stderr] --> src/numbers.rs:1:5 [INFO] [stderr] | [INFO] [stderr] 1 | use bitflags::*; [INFO] [stderr] | ^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused import: `std::borrow::Cow` [INFO] [stderr] --> src/numbers.rs:2:5 [INFO] [stderr] | [INFO] [stderr] 2 | use std::borrow::Cow; [INFO] [stderr] | ^^^^^^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused imports: `DecoderTrap`, `Encoding` [INFO] [stderr] --> src/numbers.rs:3:16 [INFO] [stderr] | [INFO] [stderr] 3 | use encoding::{Encoding, DecoderTrap}; [INFO] [stderr] | ^^^^^^^^ ^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused import: `encoding::all::UTF_8` [INFO] [stderr] --> src/numbers.rs:4:5 [INFO] [stderr] | [INFO] [stderr] 4 | use encoding::all::UTF_8; [INFO] [stderr] | ^^^^^^^^^^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused import: `Duration` [INFO] [stderr] --> src/main.rs:14:25 [INFO] [stderr] | [INFO] [stderr] 14 | use time::{PreciseTime, Duration}; [INFO] [stderr] | ^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused import: `tokenizer_peek::*` [INFO] [stderr] --> src/main.rs:16:5 [INFO] [stderr] | [INFO] [stderr] 16 | use tokenizer_peek::*; [INFO] [stderr] | ^^^^^^^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused import: `std::borrow::Cow` [INFO] [stderr] --> src/tokenizer_loop.rs:2:5 [INFO] [stderr] | [INFO] [stderr] 2 | use std::borrow::Cow; [INFO] [stderr] | ^^^^^^^^^^^^^^^^ [INFO] [stderr] | [INFO] [stderr] = note: `#[warn(unused_imports)]` on by default [INFO] [stderr] [INFO] [stderr] warning: unused imports: `DecoderTrap`, `Encoding` [INFO] [stderr] --> src/tokenizer_loop.rs:3:16 [INFO] [stderr] | [INFO] [stderr] 3 | use encoding::{Encoding, DecoderTrap}; [INFO] [stderr] | ^^^^^^^^ ^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused import: `encoding::all::UTF_8` [INFO] [stderr] --> src/tokenizer_loop.rs:4:5 [INFO] [stderr] | [INFO] [stderr] 4 | use encoding::all::UTF_8; [INFO] [stderr] | ^^^^^^^^^^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: trait objects without an explicit `dyn` are deprecated [INFO] [stderr] --> src/tokenizer_loop.rs:36:34 [INFO] [stderr] | [INFO] [stderr] 36 | pub fn words(bytes: Vec, f: &Fn(&Token) -> i32) -> Result, &'static Vec> { [INFO] [stderr] | ^^^^^^^^^^^^^^^^^ help: use `dyn`: `dyn Fn(&Token) -> i32` [INFO] [stderr] | [INFO] [stderr] = note: `#[warn(bare_trait_objects)]` on by default [INFO] [stderr] [INFO] [stderr] warning: `...` range patterns are deprecated [INFO] [stderr] --> src/tokenizer_loop.rs:69:22 [INFO] [stderr] | [INFO] [stderr] 69 | 65 ... 90 | 97 ... 122 => { [INFO] [stderr] | ^^^ help: use `..=` for an inclusive range [INFO] [stderr] | [INFO] [stderr] = note: `#[warn(ellipsis_inclusive_range_patterns)]` on by default [INFO] [stderr] [INFO] [stderr] warning: `...` range patterns are deprecated [INFO] [stderr] --> src/tokenizer_loop.rs:69:34 [INFO] [stderr] | [INFO] [stderr] 69 | 65 ... 90 | 97 ... 122 => { [INFO] [stderr] | ^^^ help: use `..=` for an inclusive range [INFO] [stderr] [INFO] [stderr] warning: `...` range patterns are deprecated [INFO] [stderr] --> src/tokenizer_loop.rs:73:22 [INFO] [stderr] | [INFO] [stderr] 73 | 33 ... 47 | 58 ... 64 | 91 ... 96 | 123 ... 126 | 160 ... 191 => { [INFO] [stderr] | ^^^ help: use `..=` for an inclusive range [INFO] [stderr] [INFO] [stderr] warning: `...` range patterns are deprecated [INFO] [stderr] --> src/tokenizer_loop.rs:73:34 [INFO] [stderr] | [INFO] [stderr] 73 | 33 ... 47 | 58 ... 64 | 91 ... 96 | 123 ... 126 | 160 ... 191 => { [INFO] [stderr] | ^^^ help: use `..=` for an inclusive range [INFO] [stderr] [INFO] [stderr] warning: `...` range patterns are deprecated [INFO] [stderr] --> src/tokenizer_loop.rs:73:46 [INFO] [stderr] | [INFO] [stderr] 73 | 33 ... 47 | 58 ... 64 | 91 ... 96 | 123 ... 126 | 160 ... 191 => { [INFO] [stderr] | ^^^ help: use `..=` for an inclusive range [INFO] [stderr] [INFO] [stderr] warning: `...` range patterns are deprecated [INFO] [stderr] --> src/tokenizer_loop.rs:73:59 [INFO] [stderr] | [INFO] [stderr] 73 | 33 ... 47 | 58 ... 64 | 91 ... 96 | 123 ... 126 | 160 ... 191 => { [INFO] [stderr] | ^^^ help: use `..=` for an inclusive range [INFO] [stderr] [INFO] [stderr] warning: `...` range patterns are deprecated [INFO] [stderr] --> src/tokenizer_loop.rs:73:73 [INFO] [stderr] | [INFO] [stderr] 73 | 33 ... 47 | 58 ... 64 | 91 ... 96 | 123 ... 126 | 160 ... 191 => { [INFO] [stderr] | ^^^ help: use `..=` for an inclusive range [INFO] [stderr] [INFO] [stderr] warning: `...` range patterns are deprecated [INFO] [stderr] --> src/tokenizer_loop.rs:77:22 [INFO] [stderr] | [INFO] [stderr] 77 | 48 ... 57 => { [INFO] [stderr] | ^^^ help: use `..=` for an inclusive range [INFO] [stderr] [INFO] [stderr] warning: `...` range patterns are deprecated [INFO] [stderr] --> src/tokenizer_loop.rs:81:23 [INFO] [stderr] | [INFO] [stderr] 81 | 127 ... 159 | 1 ... 31 => { [INFO] [stderr] | ^^^ help: use `..=` for an inclusive range [INFO] [stderr] [INFO] [stderr] warning: `...` range patterns are deprecated [INFO] [stderr] --> src/tokenizer_loop.rs:81:35 [INFO] [stderr] | [INFO] [stderr] 81 | 127 ... 159 | 1 ... 31 => { [INFO] [stderr] | ^^^ help: use `..=` for an inclusive range [INFO] [stderr] [INFO] [stderr] warning: `...` range patterns are deprecated [INFO] [stderr] --> src/tokenizer_loop.rs:85:23 [INFO] [stderr] | [INFO] [stderr] 85 | 192 ... 255 => { [INFO] [stderr] | ^^^ help: use `..=` for an inclusive range [INFO] [stderr] [INFO] [stderr] warning: unused import: `bitflags::*` [INFO] [stderr] --> src/numbers.rs:1:5 [INFO] [stderr] | [INFO] [stderr] 1 | use bitflags::*; [INFO] [stderr] | ^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused import: `std::borrow::Cow` [INFO] [stderr] --> src/numbers.rs:2:5 [INFO] [stderr] | [INFO] [stderr] 2 | use std::borrow::Cow; [INFO] [stderr] | ^^^^^^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused imports: `DecoderTrap`, `Encoding` [INFO] [stderr] --> src/numbers.rs:3:16 [INFO] [stderr] | [INFO] [stderr] 3 | use encoding::{Encoding, DecoderTrap}; [INFO] [stderr] | ^^^^^^^^ ^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused import: `encoding::all::UTF_8` [INFO] [stderr] --> src/numbers.rs:4:5 [INFO] [stderr] | [INFO] [stderr] 4 | use encoding::all::UTF_8; [INFO] [stderr] | ^^^^^^^^^^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused import: `Duration` [INFO] [stderr] --> src/main.rs:14:25 [INFO] [stderr] | [INFO] [stderr] 14 | use time::{PreciseTime, Duration}; [INFO] [stderr] | ^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused import: `tokenizer_peek::*` [INFO] [stderr] --> src/main.rs:16:5 [INFO] [stderr] | [INFO] [stderr] 16 | use tokenizer_peek::*; [INFO] [stderr] | ^^^^^^^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused variable: `token` [INFO] [stderr] --> src/tokenizer_loop.rs:97:17 [INFO] [stderr] | [INFO] [stderr] 97 | let token = Token { [INFO] [stderr] | ^^^^^ help: consider prefixing with an underscore: `_token` [INFO] [stderr] | [INFO] [stderr] = note: `#[warn(unused_variables)]` on by default [INFO] [stderr] [INFO] [stderr] warning: unused variable: `length` [INFO] [stderr] --> src/tokenizer_peek.rs:75:11 [INFO] [stderr] | [INFO] [stderr] 75 | let length = c.len(); [INFO] [stderr] | ^^^^^^ help: consider prefixing with an underscore: `_length` [INFO] [stderr] [INFO] [stderr] warning: unused variable: `i` [INFO] [stderr] --> src/tokenizer_peek.rs:77:12 [INFO] [stderr] | [INFO] [stderr] 77 | for (i, byte) in c.iter().enumerate() { [INFO] [stderr] | ^ help: consider prefixing with an underscore: `_i` [INFO] [stderr] [INFO] [stderr] warning: value assigned to `pos` is never read [INFO] [stderr] --> src/numbers.rs:11:15 [INFO] [stderr] | [INFO] [stderr] 11 | let mut pos: usize = 0; [INFO] [stderr] | ^^^ [INFO] [stderr] | [INFO] [stderr] = note: `#[warn(unused_assignments)]` on by default [INFO] [stderr] = help: maybe it is overwritten before being read? [INFO] [stderr] [INFO] [stderr] warning: variant is never constructed: `Alpha` [INFO] [stderr] --> src/tokenizer_peek.rs:16:7 [INFO] [stderr] | [INFO] [stderr] 16 | Alpha, // Pure alpha string [INFO] [stderr] | ^^^^^ [INFO] [stderr] | [INFO] [stderr] = note: `#[warn(dead_code)]` on by default [INFO] [stderr] [INFO] [stderr] warning: variant is never constructed: `Other` [INFO] [stderr] --> src/tokenizer_peek.rs:17:7 [INFO] [stderr] | [INFO] [stderr] 17 | Other, // Unidentified elements [INFO] [stderr] | ^^^^^ [INFO] [stderr] [INFO] [stderr] warning: variant is never constructed: `Numeric` [INFO] [stderr] --> src/tokenizer_peek.rs:21:7 [INFO] [stderr] | [INFO] [stderr] 21 | Numeric, // Pure numeric strings 123 [INFO] [stderr] | ^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: variant is never constructed: `NonAlphaNum` [INFO] [stderr] --> src/tokenizer_peek.rs:22:7 [INFO] [stderr] | [INFO] [stderr] 22 | NonAlphaNum // Strings with only | ¦ § _ ~ ^ [INFO] [stderr] | ^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: function `U80_FF` should have a snake case name [INFO] [stderr] --> src/tokenizer_loop.rs:141:8 [INFO] [stderr] | [INFO] [stderr] 141 | pub fn U80_FF(byte: u8) -> [u8; 2] { [INFO] [stderr] | ^^^^^^ help: convert the identifier to snake case: `u80_ff` [INFO] [stderr] | [INFO] [stderr] = note: `#[warn(non_snake_case)]` on by default [INFO] [stderr] [INFO] [stderr] warning: unused variable: `token` [INFO] [stderr] --> src/tokenizer_loop.rs:97:17 [INFO] [stderr] | [INFO] [stderr] 97 | let token = Token { [INFO] [stderr] | ^^^^^ help: consider prefixing with an underscore: `_token` [INFO] [stderr] | [INFO] [stderr] = note: `#[warn(unused_variables)]` on by default [INFO] [stderr] [INFO] [stderr] warning: unused variable: `length` [INFO] [stderr] --> src/tokenizer_peek.rs:75:11 [INFO] [stderr] | [INFO] [stderr] 75 | let length = c.len(); [INFO] [stderr] | ^^^^^^ help: consider prefixing with an underscore: `_length` [INFO] [stderr] [INFO] [stderr] warning: unused variable: `i` [INFO] [stderr] --> src/tokenizer_peek.rs:77:12 [INFO] [stderr] | [INFO] [stderr] 77 | for (i, byte) in c.iter().enumerate() { [INFO] [stderr] | ^ help: consider prefixing with an underscore: `_i` [INFO] [stderr] [INFO] [stderr] warning: value assigned to `pos` is never read [INFO] [stderr] --> src/numbers.rs:11:15 [INFO] [stderr] | [INFO] [stderr] 11 | let mut pos: usize = 0; [INFO] [stderr] | ^^^ [INFO] [stderr] | [INFO] [stderr] = note: `#[warn(unused_assignments)]` on by default [INFO] [stderr] = help: maybe it is overwritten before being read? [INFO] [stderr] [INFO] [stderr] warning: variant is never constructed: `Alpha` [INFO] [stderr] --> src/tokenizer_peek.rs:16:7 [INFO] [stderr] | [INFO] [stderr] 16 | Alpha, // Pure alpha string [INFO] [stderr] | ^^^^^ [INFO] [stderr] | [INFO] [stderr] = note: `#[warn(dead_code)]` on by default [INFO] [stderr] [INFO] [stderr] warning: variant is never constructed: `Other` [INFO] [stderr] --> src/tokenizer_peek.rs:17:7 [INFO] [stderr] | [INFO] [stderr] 17 | Other, // Unidentified elements [INFO] [stderr] | ^^^^^ [INFO] [stderr] [INFO] [stderr] warning: variant is never constructed: `Numeric` [INFO] [stderr] --> src/tokenizer_peek.rs:21:7 [INFO] [stderr] | [INFO] [stderr] 21 | Numeric, // Pure numeric strings 123 [INFO] [stderr] | ^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: variant is never constructed: `NonAlphaNum` [INFO] [stderr] --> src/tokenizer_peek.rs:22:7 [INFO] [stderr] | [INFO] [stderr] 22 | NonAlphaNum // Strings with only | ¦ § _ ~ ^ [INFO] [stderr] | ^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: function `U80_FF` should have a snake case name [INFO] [stderr] --> src/tokenizer_loop.rs:141:8 [INFO] [stderr] | [INFO] [stderr] 141 | pub fn U80_FF(byte: u8) -> [u8; 2] { [INFO] [stderr] | ^^^^^^ help: convert the identifier to snake case: `u80_ff` [INFO] [stderr] | [INFO] [stderr] = note: `#[warn(non_snake_case)]` on by default [INFO] [stderr] [INFO] [stderr] Finished dev [unoptimized + debuginfo] target(s) in 3.87s [INFO] running `"docker" "inspect" "9f9539ca46731fcdeceef45b7e291b2a471eb4fbb922a18c34734397dec07f5f"` [INFO] running `"docker" "rm" "-f" "9f9539ca46731fcdeceef45b7e291b2a471eb4fbb922a18c34734397dec07f5f"` [INFO] [stdout] 9f9539ca46731fcdeceef45b7e291b2a471eb4fbb922a18c34734397dec07f5f