[INFO] fetching crate kaffe 0.2.0...
[INFO] testing kaffe-0.2.0 against master#c90bcb9571b7aab0d8beaa2ce8a998ffaf079d38 for pr-146098-7
[INFO] extracting crate kaffe 0.2.0 into /workspace/builds/worker-5-tc1/source
[INFO] started tweaking crates.io crate kaffe 0.2.0
[INFO] finished tweaking crates.io crate kaffe 0.2.0
[INFO] tweaked toml for crates.io crate kaffe 0.2.0 written to /workspace/builds/worker-5-tc1/source/Cargo.toml
[INFO] validating manifest of crates.io crate kaffe 0.2.0 on toolchain c90bcb9571b7aab0d8beaa2ce8a998ffaf079d38
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+c90bcb9571b7aab0d8beaa2ce8a998ffaf079d38" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }`
[INFO] crate crates.io crate kaffe 0.2.0 already has a lockfile, it will not be regenerated
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+c90bcb9571b7aab0d8beaa2ce8a998ffaf079d38" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }`
[INFO] [stderr]     Updating crates.io index
[INFO] [stderr]  Downloading crates ...
[INFO] [stderr]   Downloaded serde_derive v1.0.164
[INFO] [stderr]   Downloaded memoffset v0.8.0
[INFO] [stderr]   Downloaded errno-dragonfly v0.1.2
[INFO] [stderr]   Downloaded anstyle v1.0.0
[INFO] [stderr]   Downloaded hermit-abi v0.3.1
[INFO] [stderr]   Downloaded ciborium-io v0.2.1
[INFO] [stderr]   Downloaded ciborium-ll v0.2.1
[INFO] [stderr]   Downloaded anyhow v1.0.71
[INFO] [stderr]   Downloaded half v1.8.2
[INFO] [stderr]   Downloaded cc v1.0.79
[INFO] [stderr]   Downloaded js-sys v0.3.63
[INFO] [stderr]   Downloaded bumpalo v3.13.0
[INFO] [stderr]   Downloaded wasm-bindgen-backend v0.2.86
[INFO] [stderr]   Downloaded crossbeam-channel v0.5.8
[INFO] [stderr]   Downloaded wasm-bindgen v0.2.86
[INFO] [stderr]   Downloaded clap_builder v4.3.3
[INFO] [stderr]   Downloaded serde_json v1.0.96
[INFO] [stderr]   Downloaded syn v2.0.18
[INFO] [stderr]   Downloaded plotters v0.3.4
[INFO] [stderr]   Downloaded rayon v1.7.0
[INFO] [stderr]   Downloaded regex v1.8.4
[INFO] [stderr]   Downloaded ryu v1.0.13
[INFO] [stderr]   Downloaded crossbeam-utils v0.8.15
[INFO] [stderr]   Downloaded rustix v0.37.19
[INFO] [stderr]   Downloaded windows_aarch64_gnullvm v0.48.0
[INFO] [stderr]   Downloaded windows_x86_64_gnullvm v0.48.0
[INFO] [stderr]   Downloaded regex-syntax v0.7.2
[INFO] [stderr]   Downloaded log v0.4.19
[INFO] [stderr]   Downloaded once_cell v1.18.0
[INFO] [stderr]   Downloaded crossbeam-epoch v0.9.14
[INFO] [stderr]   Downloaded crossbeam-deque v0.8.3
[INFO] [stderr]   Downloaded rayon-core v1.11.0
[INFO] [stderr]   Downloaded proc-macro2 v1.0.60
[INFO] [stderr]   Downloaded serde v1.0.164
[INFO] [stderr]   Downloaded ciborium v0.2.1
[INFO] [stderr]   Downloaded clap v4.3.3
[INFO] [stderr]   Downloaded plotters-backend v0.3.4
[INFO] [stderr]   Downloaded wasm-bindgen-macro-support v0.2.86
[INFO] [stderr]   Downloaded itoa v1.0.6
[INFO] [stderr]   Downloaded ppv-lite86 v0.2.17
[INFO] [stderr]   Downloaded windows_aarch64_msvc v0.48.0
[INFO] [stderr]   Downloaded windows_x86_64_msvc v0.48.0
[INFO] [stderr]   Downloaded unicode-ident v1.0.9
[INFO] [stderr]   Downloaded windows_x86_64_gnu v0.48.0
[INFO] [stderr]   Downloaded io-lifetimes v1.0.11
[INFO] [stderr]   Downloaded windows_i686_msvc v0.48.0
[INFO] [stderr]   Downloaded web-sys v0.3.63
[INFO] [stderr]   Downloaded windows_i686_gnu v0.48.0
[INFO] [stderr]   Downloaded libc v0.2.146
[INFO] [stderr]   Downloaded oorandom v11.1.3
[INFO] [stderr]   Downloaded wasm-bindgen-shared v0.2.86
[INFO] [stderr]   Downloaded getrandom v0.2.10
[INFO] [stderr]   Downloaded windows-targets v0.48.0
[INFO] [stderr]   Downloaded wasm-bindgen-macro v0.2.86
[INFO] [stderr]   Downloaded walkdir v2.3.3
[INFO] [stderr]   Downloaded either v1.8.1
[INFO] [stderr]   Downloaded linux-raw-sys v0.3.8
[INFO] [stderr]   Downloaded num_cpus v1.15.0
[INFO] [stderr]   Downloaded quote v1.0.28
[INFO] [stderr]   Downloaded clap_lex v0.5.0
[INFO] [stderr]   Downloaded is-terminal v0.4.7
[INFO] [stderr]   Downloaded plotters-svg v0.3.3
[INFO] [stderr]   Downloaded errno v0.3.1
[INFO] [stderr]   Downloaded hermit-abi v0.2.6
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:4848fb76d95f26979359cc7e45710b1dbc8f3acb7aeedee7c460d7702230f228" "/opt/rustwide/cargo-home/bin/cargo" "+c90bcb9571b7aab0d8beaa2ce8a998ffaf079d38" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }`
[INFO] [stdout] bde79f39c0026cbdb0cf6bfc668e8db45a8f35eb5ed5b613f9f04a38372158b0
[INFO] running `Command { std: "docker" "start" "-a" "bde79f39c0026cbdb0cf6bfc668e8db45a8f35eb5ed5b613f9f04a38372158b0", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "inspect" "bde79f39c0026cbdb0cf6bfc668e8db45a8f35eb5ed5b613f9f04a38372158b0", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "bde79f39c0026cbdb0cf6bfc668e8db45a8f35eb5ed5b613f9f04a38372158b0", kill_on_drop: false }`
[INFO] [stdout] bde79f39c0026cbdb0cf6bfc668e8db45a8f35eb5ed5b613f9f04a38372158b0
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:4848fb76d95f26979359cc7e45710b1dbc8f3acb7aeedee7c460d7702230f228" "/opt/rustwide/cargo-home/bin/cargo" "+c90bcb9571b7aab0d8beaa2ce8a998ffaf079d38" "build" "--frozen" "--message-format=json", kill_on_drop: false }`
[INFO] [stdout] 6bcd7f1e7166809a5d07478a871fa73402cd985c4bd917795fbe231bd3fdd23d
[INFO] running `Command { std: "docker" "start" "-a" "6bcd7f1e7166809a5d07478a871fa73402cd985c4bd917795fbe231bd3fdd23d", kill_on_drop: false }`
[INFO] [stderr]    Compiling autocfg v1.1.0
[INFO] [stderr]    Compiling libc v0.2.146
[INFO] [stderr]    Compiling crossbeam-utils v0.8.15
[INFO] [stderr]    Compiling proc-macro2 v1.0.60
[INFO] [stderr]    Compiling unicode-ident v1.0.9
[INFO] [stderr]    Compiling quote v1.0.28
[INFO] [stderr]    Compiling scopeguard v1.1.0
[INFO] [stderr]    Compiling rayon-core v1.11.0
[INFO] [stderr]    Compiling anyhow v1.0.71
[INFO] [stderr]    Compiling serde v1.0.164
[INFO] [stderr]    Compiling ppv-lite86 v0.2.17
[INFO] [stderr]    Compiling either v1.8.1
[INFO] [stderr]    Compiling itertools v0.10.5
[INFO] [stderr]    Compiling memoffset v0.8.0
[INFO] [stderr]    Compiling crossbeam-epoch v0.9.14
[INFO] [stderr]    Compiling num-traits v0.2.15
[INFO] [stderr]    Compiling crossbeam-channel v0.5.8
[INFO] [stderr]    Compiling syn v2.0.18
[INFO] [stderr]    Compiling getrandom v0.2.10
[INFO] [stderr]    Compiling num_cpus v1.15.0
[INFO] [stderr]    Compiling rand_core v0.6.4
[INFO] [stderr]    Compiling crossbeam-deque v0.8.3
[INFO] [stderr]    Compiling rand_chacha v0.3.1
[INFO] [stderr]    Compiling rand v0.8.5
[INFO] [stderr]    Compiling rayon v1.7.0
[INFO] [stderr]    Compiling serde_derive v1.0.164
[INFO] [stderr]    Compiling kaffe v0.2.0 (/opt/rustwide/workdir)
[INFO] [stdout] warning: unused import: `marker::PhantomData`
[INFO] [stdout]  --> src/nn/optimizer.rs:4:25
[INFO] [stdout]   |
[INFO] [stdout] 4 | use std::{error::Error, marker::PhantomData, str::FromStr};
[INFO] [stdout]   |                         ^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout]   = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `rayon::prelude::*`
[INFO] [stdout]  --> src/nn/optimizer.rs:7:5
[INFO] [stdout]   |
[INFO] [stdout] 7 | use rayon::prelude::*;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `idx`
[INFO] [stdout]   --> src/nn/dataset.rs:57:26
[INFO] [stdout]    |
[INFO] [stdout] 57 |     fn get<'a, T>(&self, idx: usize) -> (Tensor<'a, T>, String)
[INFO] [stdout]    |                          ^^^ help: if this is intentional, prefix it with an underscore: `_idx`
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(unused_variables)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y`
[INFO] [stdout]   --> src/nn/loss.rs:31:30
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                              ^ help: if this is intentional, prefix it with an underscore: `_y`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y_hat`
[INFO] [stdout]   --> src/nn/loss.rs:31:49
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                                                 ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y`
[INFO] [stdout]   --> src/nn/loss.rs:50:29
[INFO] [stdout]    |
[INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                             ^ help: if this is intentional, prefix it with an underscore: `_y`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y_hat`
[INFO] [stdout]   --> src/nn/loss.rs:50:48
[INFO] [stdout]    |
[INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                                                ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `cost`
[INFO] [stdout]   --> src/nn/optimizer.rs:75:31
[INFO] [stdout]    |
[INFO] [stdout] 75 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]    |                               ^^^^ help: if this is intentional, prefix it with an underscore: `_cost`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `vars`
[INFO] [stdout]   --> src/nn/optimizer.rs:75:40
[INFO] [stdout]    |
[INFO] [stdout] 75 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]    |                                        ^^^^ help: if this is intentional, prefix it with an underscore: `_vars`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `cost`
[INFO] [stdout]    --> src/nn/optimizer.rs:117:31
[INFO] [stdout]     |
[INFO] [stdout] 117 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]     |                               ^^^^ help: if this is intentional, prefix it with an underscore: `_cost`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `vars`
[INFO] [stdout]    --> src/nn/optimizer.rs:117:40
[INFO] [stdout]     |
[INFO] [stdout] 117 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]     |                                        ^^^^ help: if this is intentional, prefix it with an underscore: `_vars`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `tensor`
[INFO] [stdout]   --> src/nn/mod.rs:24:5
[INFO] [stdout]    |
[INFO] [stdout] 24 |     tensor: Tensor<'a, T>,
[INFO] [stdout]    |     ^^^^^^ help: if this is intentional, prefix it with an underscore: `_tensor`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `kernel`
[INFO] [stdout]   --> src/nn/mod.rs:25:5
[INFO] [stdout]    |
[INFO] [stdout] 25 |     kernel: Tensor<'a, T>,
[INFO] [stdout]    |     ^^^^^^ help: if this is intentional, prefix it with an underscore: `_kernel`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `stride`
[INFO] [stdout]   --> src/nn/mod.rs:26:5
[INFO] [stdout]    |
[INFO] [stdout] 26 |     stride: usize,
[INFO] [stdout]    |     ^^^^^^ help: if this is intentional, prefix it with an underscore: `_stride`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `padding`
[INFO] [stdout]   --> src/nn/mod.rs:27:5
[INFO] [stdout]    |
[INFO] [stdout] 27 |     padding: usize,
[INFO] [stdout]    |     ^^^^^^^ help: if this is intentional, prefix it with an underscore: `_padding`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `f`
[INFO] [stdout]    --> src/tensor/mod.rs:257:19
[INFO] [stdout]     |
[INFO] [stdout] 257 |     fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
[INFO] [stdout]     |                   ^ help: if this is intentional, prefix it with an underscore: `_f`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `start_idx`
[INFO] [stdout]    --> src/tensor/mod.rs:686:30
[INFO] [stdout]     |
[INFO] [stdout] 686 |     fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec<T> {
[INFO] [stdout]     |                              ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_start_idx`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `size`
[INFO] [stdout]    --> src/tensor/mod.rs:686:48
[INFO] [stdout]     |
[INFO] [stdout] 686 |     fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec<T> {
[INFO] [stdout]     |                                                ^^^^ help: if this is intentional, prefix it with an underscore: `_size`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `rowcol`
[INFO] [stdout]    --> src/tensor/mod.rs:889:19
[INFO] [stdout]     |
[INFO] [stdout] 889 |     fn sum(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                   ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `dimension`
[INFO] [stdout]    --> src/tensor/mod.rs:889:34
[INFO] [stdout]     |
[INFO] [stdout] 889 |     fn sum(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                                  ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `rowcol`
[INFO] [stdout]    --> src/tensor/mod.rs:904:20
[INFO] [stdout]     |
[INFO] [stdout] 904 |     fn prod(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                    ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `dimension`
[INFO] [stdout]    --> src/tensor/mod.rs:904:35
[INFO] [stdout]     |
[INFO] [stdout] 904 |     fn prod(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                                   ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: field `layers` is never read
[INFO] [stdout]   --> src/nn/mod.rs:39:5
[INFO] [stdout]    |
[INFO] [stdout] 37 | pub struct Net {
[INFO] [stdout]    |            --- field in this struct
[INFO] [stdout] 38 |     /// Represents all layers in the network
[INFO] [stdout] 39 |     layers: Vec<Layer>,
[INFO] [stdout]    |     ^^^^^^
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: fields `data`, `batch_size`, and `shuffle` are never read
[INFO] [stdout]   --> src/nn/dataloader.rs:10:5
[INFO] [stdout]    |
[INFO] [stdout]  8 | pub struct DataLoader<D: DataSet + Clone> {
[INFO] [stdout]    |            ---------- fields in this struct
[INFO] [stdout]  9 |     /// Dataset
[INFO] [stdout] 10 |     data: D,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 11 |     /// Batch size
[INFO] [stdout] 12 |     batch_size: usize,
[INFO] [stdout]    |     ^^^^^^^^^^
[INFO] [stdout] 13 |     /// Whether or not to shuffle data
[INFO] [stdout] 14 |     shuffle: bool,
[INFO] [stdout]    |     ^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout]    = note: `DataLoader` has a derived impl for the trait `Clone`, but this is intentionally ignored during dead code analysis
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated items `new` and `iter` are never used
[INFO] [stdout]   --> src/nn/dataloader.rs:41:8
[INFO] [stdout]    |
[INFO] [stdout] 33 | impl<D: DataSet + Clone> DataLoader<D> {
[INFO] [stdout]    | -------------------------------------- associated items in this implementation
[INFO] [stdout] ...
[INFO] [stdout] 41 |     fn new(data: D, batch_size: usize, shuffle: bool) -> DataLoader<D> {
[INFO] [stdout]    |        ^^^
[INFO] [stdout] ...
[INFO] [stdout] 56 |     fn iter(&mut self) -> DataLoader<D> {
[INFO] [stdout]    |        ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: fields `img_dir`, `transform`, and `target_transform` are never read
[INFO] [stdout]   --> src/nn/dataset.rs:33:5
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub struct CustomImageDataset {
[INFO] [stdout]    |            ------------------ fields in this struct
[INFO] [stdout] 32 |     img_labels: Vec<String>,
[INFO] [stdout] 33 |     img_dir: &'static str,
[INFO] [stdout]    |     ^^^^^^^
[INFO] [stdout] 34 |     transform: Option<Transform>,
[INFO] [stdout]    |     ^^^^^^^^^
[INFO] [stdout] 35 |     target_transform: Option<Transform>,
[INFO] [stdout]    |     ^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: multiple fields are never read
[INFO] [stdout]   --> src/nn/optimizer.rs:38:5
[INFO] [stdout]    |
[INFO] [stdout] 36 | pub struct Adam<T> {
[INFO] [stdout]    |            ---- fields in this struct
[INFO] [stdout] 37 |     /// Learning rate
[INFO] [stdout] 38 |     lr: T,
[INFO] [stdout]    |     ^^
[INFO] [stdout] 39 |     /// Momentum
[INFO] [stdout] 40 |     momentum: f32,
[INFO] [stdout]    |     ^^^^^^^^
[INFO] [stdout] 41 |     /// Decay rate
[INFO] [stdout] 42 |     decay_rate: f32,
[INFO] [stdout]    |     ^^^^^^^^^^
[INFO] [stdout] 43 |     /// Beta 1
[INFO] [stdout] 44 |     beta1: f32,
[INFO] [stdout]    |     ^^^^^
[INFO] [stdout] 45 |     /// Beta 2:
[INFO] [stdout] 46 |     beta2: f32,
[INFO] [stdout]    |     ^^^^^
[INFO] [stdout] 47 |     /// Epsilon
[INFO] [stdout] 48 |     epsilon: f32,
[INFO] [stdout]    |     ^^^^^^^
[INFO] [stdout] 49 |     m_dw: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 50 |     v_dw: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 51 |     m_db: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 52 |     v_db: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: fields `lr`, `momentum`, and `decay_rate` are never read
[INFO] [stdout]    --> src/nn/optimizer.rs:97:5
[INFO] [stdout]     |
[INFO] [stdout]  95 | pub struct SGD<T> {
[INFO] [stdout]     |            --- fields in this struct
[INFO] [stdout]  96 |     /// Learning rate
[INFO] [stdout]  97 |     lr: T,
[INFO] [stdout]     |     ^^
[INFO] [stdout]  98 |     /// Momentum
[INFO] [stdout]  99 |     momentum: f32,
[INFO] [stdout]     |     ^^^^^^^^
[INFO] [stdout] 100 |     /// Decay rate
[INFO] [stdout] 101 |     decay_rate: f32,
[INFO] [stdout]     |     ^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `swap` is never used
[INFO] [stdout]   --> src/tensor/mod.rs:33:4
[INFO] [stdout]    |
[INFO] [stdout] 33 | fn swap(lhs: &mut usize, rhs: &mut usize) {
[INFO] [stdout]    |    ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: methods `get_sub_tensor`, `sum`, and `prod` are never used
[INFO] [stdout]    --> src/tensor/mod.rs:686:8
[INFO] [stdout]     |
[INFO] [stdout] 560 | / impl<'a, T> Tensor<'a, T>
[INFO] [stdout] 561 | | where
[INFO] [stdout] 562 | |     T: TensorElement + Div<Output = T> + Sum<T>,
[INFO] [stdout] 563 | |     <T as FromStr>::Err: Error + 'static,
[INFO] [stdout] 564 | |     Vec<T>: IntoParallelIterator,
[INFO] [stdout] 565 | |     Vec<&'a T>: IntoParallelRefIterator<'a>,
[INFO] [stdout]     | |____________________________________________- methods in this implementation
[INFO] [stdout] ...
[INFO] [stdout] 686 |       fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec<T> {
[INFO] [stdout]     |          ^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout] 889 |       fn sum(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |          ^^^
[INFO] [stdout] ...
[INFO] [stdout] 904 |       fn prod(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |          ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `ReLU` should have a snake case name
[INFO] [stdout]   --> src/nn/activation.rs:25:8
[INFO] [stdout]    |
[INFO] [stdout] 25 | pub fn ReLU<'a, 'b, T>(x: &Tensor<'a, T>) -> Tensor<'b, T>
[INFO] [stdout]    |        ^^^^ help: convert the identifier to snake case: `re_lu`
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(non_snake_case)]` (part of `#[warn(nonstandard_style)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `PReLU` should have a snake case name
[INFO] [stdout]   --> src/nn/activation.rs:52:8
[INFO] [stdout]    |
[INFO] [stdout] 52 | pub fn PReLU<'a, 'b, T>(x: &Tensor<'a, T>, alpha: T) -> Tensor<'b, T>
[INFO] [stdout]    |        ^^^^^ help: convert the identifier to snake case: `pre_lu`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `Sigmoid` should have a snake case name
[INFO] [stdout]   --> src/nn/activation.rs:79:8
[INFO] [stdout]    |
[INFO] [stdout] 79 | pub fn Sigmoid<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case (notice the capitalization): `sigmoid`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `GeLU` should have a snake case name
[INFO] [stdout]    --> src/nn/activation.rs:108:8
[INFO] [stdout]     |
[INFO] [stdout] 108 | pub fn GeLU<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T>
[INFO] [stdout]     |        ^^^^ help: convert the identifier to snake case: `ge_lu`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: missing documentation for an associated function
[INFO] [stdout]   --> src/nn/loss.rs:14:5
[INFO] [stdout]    |
[INFO] [stdout] 14 |     fn loss(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T;
[INFO] [stdout]    |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout] note: the lint level is defined here
[INFO] [stdout]   --> src/nn/loss.rs:2:9
[INFO] [stdout]    |
[INFO] [stdout]  2 | #![warn(missing_docs)]
[INFO] [stdout]    |         ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: missing documentation for an associated function
[INFO] [stdout]   --> src/nn/loss.rs:15:5
[INFO] [stdout]    |
[INFO] [stdout] 15 |     fn backward();
[INFO] [stdout]    |     ^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `BCEntroypyLoss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:31:8
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^^^^^^^^^ help: convert the identifier to snake case: `bcentroypy_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `CEntroypyLoss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:50:8
[INFO] [stdout]    |
[INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^^^^^^^^ help: convert the identifier to snake case: `centroypy_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `L1Loss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:73:8
[INFO] [stdout]    |
[INFO] [stdout] 73 | pub fn L1Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^ help: convert the identifier to snake case: `l1_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `L2Loss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:96:8
[INFO] [stdout]    |
[INFO] [stdout] 96 | pub fn L2Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^ help: convert the identifier to snake case: `l2_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `MaxPool` should have a snake case name
[INFO] [stdout]   --> src/nn/pooling.rs:23:8
[INFO] [stdout]    |
[INFO] [stdout] 23 | pub fn MaxPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case: `max_pool`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `MinPool` should have a snake case name
[INFO] [stdout]   --> src/nn/pooling.rs:62:8
[INFO] [stdout]    |
[INFO] [stdout] 62 | pub fn MinPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case: `min_pool`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `AvgPool` should have a snake case name
[INFO] [stdout]   --> src/nn/pooling.rs:93:8
[INFO] [stdout]    |
[INFO] [stdout] 93 | pub fn AvgPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case: `avg_pool`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: missing documentation for a struct
[INFO] [stdout]  --> src/nn/transform.rs:4:1
[INFO] [stdout]   |
[INFO] [stdout] 4 | pub struct Transform;
[INFO] [stdout]   | ^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout] note: the lint level is defined here
[INFO] [stdout]  --> src/nn/transform.rs:2:9
[INFO] [stdout]   |
[INFO] [stdout] 2 | #![warn(missing_docs)]
[INFO] [stdout]   |         ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: missing documentation for a struct
[INFO] [stdout]    --> src/tensor/mod.rs:145:1
[INFO] [stdout]     |
[INFO] [stdout] 145 | pub struct Tensor<'a, T>
[INFO] [stdout]     | ^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] note: the lint level is defined here
[INFO] [stdout]    --> src/lib.rs:1:9
[INFO] [stdout]     |
[INFO] [stdout]   1 | #![warn(missing_docs)]
[INFO] [stdout]     |         ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stderr]     Finished `dev` profile [unoptimized + debuginfo] target(s) in 14.76s
[INFO] running `Command { std: "docker" "inspect" "6bcd7f1e7166809a5d07478a871fa73402cd985c4bd917795fbe231bd3fdd23d", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "6bcd7f1e7166809a5d07478a871fa73402cd985c4bd917795fbe231bd3fdd23d", kill_on_drop: false }`
[INFO] [stdout] 6bcd7f1e7166809a5d07478a871fa73402cd985c4bd917795fbe231bd3fdd23d
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:4848fb76d95f26979359cc7e45710b1dbc8f3acb7aeedee7c460d7702230f228" "/opt/rustwide/cargo-home/bin/cargo" "+c90bcb9571b7aab0d8beaa2ce8a998ffaf079d38" "test" "--frozen" "--no-run" "--message-format=json", kill_on_drop: false }`
[INFO] [stdout] 25623b942f78b7a8ad9c6331268b5e7e4728d9bb210f85be405d68f5c1a67c7f
[INFO] running `Command { std: "docker" "start" "-a" "25623b942f78b7a8ad9c6331268b5e7e4728d9bb210f85be405d68f5c1a67c7f", kill_on_drop: false }`
[INFO] [stderr]    Compiling libc v0.2.146
[INFO] [stderr]    Compiling serde v1.0.164
[INFO] [stderr]    Compiling io-lifetimes v1.0.11
[INFO] [stderr]    Compiling bitflags v1.3.2
[INFO] [stderr]    Compiling rustix v0.37.19
[INFO] [stderr]    Compiling serde_json v1.0.96
[INFO] [stderr]    Compiling ryu v1.0.13
[INFO] [stderr]    Compiling anstyle v1.0.0
[INFO] [stderr]    Compiling half v1.8.2
[INFO] [stderr]    Compiling plotters-backend v0.3.4
[INFO] [stderr]    Compiling ciborium-io v0.2.1
[INFO] [stderr]    Compiling linux-raw-sys v0.3.8
[INFO] [stderr]    Compiling clap_lex v0.5.0
[INFO] [stderr]    Compiling itoa v1.0.6
[INFO] [stderr]    Compiling same-file v1.0.6
[INFO] [stderr]    Compiling cast v0.3.0
[INFO] [stderr]    Compiling regex-syntax v0.7.2
[INFO] [stderr]    Compiling anes v0.1.6
[INFO] [stderr]    Compiling criterion-plot v0.5.0
[INFO] [stderr]    Compiling clap_builder v4.3.3
[INFO] [stderr]    Compiling plotters-svg v0.3.3
[INFO] [stderr]    Compiling ciborium-ll v0.2.1
[INFO] [stderr]    Compiling walkdir v2.3.3
[INFO] [stderr]    Compiling oorandom v11.1.3
[INFO] [stderr]    Compiling once_cell v1.18.0
[INFO] [stderr]    Compiling plotters v0.3.4
[INFO] [stderr]    Compiling getrandom v0.2.10
[INFO] [stderr]    Compiling num_cpus v1.15.0
[INFO] [stderr]    Compiling regex v1.8.4
[INFO] [stderr]    Compiling rayon-core v1.11.0
[INFO] [stderr]    Compiling rand_core v0.6.4
[INFO] [stderr]    Compiling rand_chacha v0.3.1
[INFO] [stderr]    Compiling rand v0.8.5
[INFO] [stderr]    Compiling rayon v1.7.0
[INFO] [stderr]    Compiling clap v4.3.3
[INFO] [stderr]    Compiling ciborium v0.2.1
[INFO] [stderr]    Compiling is-terminal v0.4.7
[INFO] [stderr]    Compiling tinytemplate v1.2.1
[INFO] [stderr]    Compiling criterion v0.5.1
[INFO] [stderr]    Compiling kaffe v0.2.0 (/opt/rustwide/workdir)
[INFO] [stdout] warning: unused import: `marker::PhantomData`
[INFO] [stdout]  --> src/nn/optimizer.rs:4:25
[INFO] [stdout]   |
[INFO] [stdout] 4 | use std::{error::Error, marker::PhantomData, str::FromStr};
[INFO] [stdout]   |                         ^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout]   = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `rayon::prelude::*`
[INFO] [stdout]  --> src/nn/optimizer.rs:7:5
[INFO] [stdout]   |
[INFO] [stdout] 7 | use rayon::prelude::*;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `idx`
[INFO] [stdout]   --> src/nn/dataset.rs:57:26
[INFO] [stdout]    |
[INFO] [stdout] 57 |     fn get<'a, T>(&self, idx: usize) -> (Tensor<'a, T>, String)
[INFO] [stdout]    |                          ^^^ help: if this is intentional, prefix it with an underscore: `_idx`
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(unused_variables)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y`
[INFO] [stdout]   --> src/nn/loss.rs:31:30
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                              ^ help: if this is intentional, prefix it with an underscore: `_y`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y_hat`
[INFO] [stdout]   --> src/nn/loss.rs:31:49
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                                                 ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y`
[INFO] [stdout]   --> src/nn/loss.rs:50:29
[INFO] [stdout]    |
[INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                             ^ help: if this is intentional, prefix it with an underscore: `_y`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y_hat`
[INFO] [stdout]   --> src/nn/loss.rs:50:48
[INFO] [stdout]    |
[INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                                                ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `cost`
[INFO] [stdout]   --> src/nn/optimizer.rs:75:31
[INFO] [stdout]    |
[INFO] [stdout] 75 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]    |                               ^^^^ help: if this is intentional, prefix it with an underscore: `_cost`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `vars`
[INFO] [stdout]   --> src/nn/optimizer.rs:75:40
[INFO] [stdout]    |
[INFO] [stdout] 75 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]    |                                        ^^^^ help: if this is intentional, prefix it with an underscore: `_vars`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `cost`
[INFO] [stdout]    --> src/nn/optimizer.rs:117:31
[INFO] [stdout]     |
[INFO] [stdout] 117 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]     |                               ^^^^ help: if this is intentional, prefix it with an underscore: `_cost`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `vars`
[INFO] [stdout]    --> src/nn/optimizer.rs:117:40
[INFO] [stdout]     |
[INFO] [stdout] 117 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]     |                                        ^^^^ help: if this is intentional, prefix it with an underscore: `_vars`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `tensor`
[INFO] [stdout]   --> src/nn/mod.rs:24:5
[INFO] [stdout]    |
[INFO] [stdout] 24 |     tensor: Tensor<'a, T>,
[INFO] [stdout]    |     ^^^^^^ help: if this is intentional, prefix it with an underscore: `_tensor`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `kernel`
[INFO] [stdout]   --> src/nn/mod.rs:25:5
[INFO] [stdout]    |
[INFO] [stdout] 25 |     kernel: Tensor<'a, T>,
[INFO] [stdout]    |     ^^^^^^ help: if this is intentional, prefix it with an underscore: `_kernel`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `stride`
[INFO] [stdout]   --> src/nn/mod.rs:26:5
[INFO] [stdout]    |
[INFO] [stdout] 26 |     stride: usize,
[INFO] [stdout]    |     ^^^^^^ help: if this is intentional, prefix it with an underscore: `_stride`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `padding`
[INFO] [stdout]   --> src/nn/mod.rs:27:5
[INFO] [stdout]    |
[INFO] [stdout] 27 |     padding: usize,
[INFO] [stdout]    |     ^^^^^^^ help: if this is intentional, prefix it with an underscore: `_padding`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `f`
[INFO] [stdout]    --> src/tensor/mod.rs:257:19
[INFO] [stdout]     |
[INFO] [stdout] 257 |     fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
[INFO] [stdout]     |                   ^ help: if this is intentional, prefix it with an underscore: `_f`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `start_idx`
[INFO] [stdout]    --> src/tensor/mod.rs:686:30
[INFO] [stdout]     |
[INFO] [stdout] 686 |     fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec<T> {
[INFO] [stdout]     |                              ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_start_idx`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `size`
[INFO] [stdout]    --> src/tensor/mod.rs:686:48
[INFO] [stdout]     |
[INFO] [stdout] 686 |     fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec<T> {
[INFO] [stdout]     |                                                ^^^^ help: if this is intentional, prefix it with an underscore: `_size`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `rowcol`
[INFO] [stdout]    --> src/tensor/mod.rs:889:19
[INFO] [stdout]     |
[INFO] [stdout] 889 |     fn sum(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                   ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `dimension`
[INFO] [stdout]    --> src/tensor/mod.rs:889:34
[INFO] [stdout]     |
[INFO] [stdout] 889 |     fn sum(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                                  ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `rowcol`
[INFO] [stdout]    --> src/tensor/mod.rs:904:20
[INFO] [stdout]     |
[INFO] [stdout] 904 |     fn prod(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                    ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `dimension`
[INFO] [stdout]    --> src/tensor/mod.rs:904:35
[INFO] [stdout]     |
[INFO] [stdout] 904 |     fn prod(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                                   ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: field `layers` is never read
[INFO] [stdout]   --> src/nn/mod.rs:39:5
[INFO] [stdout]    |
[INFO] [stdout] 37 | pub struct Net {
[INFO] [stdout]    |            --- field in this struct
[INFO] [stdout] 38 |     /// Represents all layers in the network
[INFO] [stdout] 39 |     layers: Vec<Layer>,
[INFO] [stdout]    |     ^^^^^^
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: fields `data`, `batch_size`, and `shuffle` are never read
[INFO] [stdout]   --> src/nn/dataloader.rs:10:5
[INFO] [stdout]    |
[INFO] [stdout]  8 | pub struct DataLoader<D: DataSet + Clone> {
[INFO] [stdout]    |            ---------- fields in this struct
[INFO] [stdout]  9 |     /// Dataset
[INFO] [stdout] 10 |     data: D,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 11 |     /// Batch size
[INFO] [stdout] 12 |     batch_size: usize,
[INFO] [stdout]    |     ^^^^^^^^^^
[INFO] [stdout] 13 |     /// Whether or not to shuffle data
[INFO] [stdout] 14 |     shuffle: bool,
[INFO] [stdout]    |     ^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout]    = note: `DataLoader` has a derived impl for the trait `Clone`, but this is intentionally ignored during dead code analysis
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated items `new` and `iter` are never used
[INFO] [stdout]   --> src/nn/dataloader.rs:41:8
[INFO] [stdout]    |
[INFO] [stdout] 33 | impl<D: DataSet + Clone> DataLoader<D> {
[INFO] [stdout]    | -------------------------------------- associated items in this implementation
[INFO] [stdout] ...
[INFO] [stdout] 41 |     fn new(data: D, batch_size: usize, shuffle: bool) -> DataLoader<D> {
[INFO] [stdout]    |        ^^^
[INFO] [stdout] ...
[INFO] [stdout] 56 |     fn iter(&mut self) -> DataLoader<D> {
[INFO] [stdout]    |        ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: fields `img_dir`, `transform`, and `target_transform` are never read
[INFO] [stdout]   --> src/nn/dataset.rs:33:5
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub struct CustomImageDataset {
[INFO] [stdout]    |            ------------------ fields in this struct
[INFO] [stdout] 32 |     img_labels: Vec<String>,
[INFO] [stdout] 33 |     img_dir: &'static str,
[INFO] [stdout]    |     ^^^^^^^
[INFO] [stdout] 34 |     transform: Option<Transform>,
[INFO] [stdout]    |     ^^^^^^^^^
[INFO] [stdout] 35 |     target_transform: Option<Transform>,
[INFO] [stdout]    |     ^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: multiple fields are never read
[INFO] [stdout]   --> src/nn/optimizer.rs:38:5
[INFO] [stdout]    |
[INFO] [stdout] 36 | pub struct Adam<T> {
[INFO] [stdout]    |            ---- fields in this struct
[INFO] [stdout] 37 |     /// Learning rate
[INFO] [stdout] 38 |     lr: T,
[INFO] [stdout]    |     ^^
[INFO] [stdout] 39 |     /// Momentum
[INFO] [stdout] 40 |     momentum: f32,
[INFO] [stdout]    |     ^^^^^^^^
[INFO] [stdout] 41 |     /// Decay rate
[INFO] [stdout] 42 |     decay_rate: f32,
[INFO] [stdout]    |     ^^^^^^^^^^
[INFO] [stdout] 43 |     /// Beta 1
[INFO] [stdout] 44 |     beta1: f32,
[INFO] [stdout]    |     ^^^^^
[INFO] [stdout] 45 |     /// Beta 2:
[INFO] [stdout] 46 |     beta2: f32,
[INFO] [stdout]    |     ^^^^^
[INFO] [stdout] 47 |     /// Epsilon
[INFO] [stdout] 48 |     epsilon: f32,
[INFO] [stdout]    |     ^^^^^^^
[INFO] [stdout] 49 |     m_dw: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 50 |     v_dw: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 51 |     m_db: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 52 |     v_db: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: fields `lr`, `momentum`, and `decay_rate` are never read
[INFO] [stdout]    --> src/nn/optimizer.rs:97:5
[INFO] [stdout]     |
[INFO] [stdout]  95 | pub struct SGD<T> {
[INFO] [stdout]     |            --- fields in this struct
[INFO] [stdout]  96 |     /// Learning rate
[INFO] [stdout]  97 |     lr: T,
[INFO] [stdout]     |     ^^
[INFO] [stdout]  98 |     /// Momentum
[INFO] [stdout]  99 |     momentum: f32,
[INFO] [stdout]     |     ^^^^^^^^
[INFO] [stdout] 100 |     /// Decay rate
[INFO] [stdout] 101 |     decay_rate: f32,
[INFO] [stdout]     |     ^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `swap` is never used
[INFO] [stdout]   --> src/tensor/mod.rs:33:4
[INFO] [stdout]    |
[INFO] [stdout] 33 | fn swap(lhs: &mut usize, rhs: &mut usize) {
[INFO] [stdout]    |    ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: methods `get_sub_tensor`, `sum`, and `prod` are never used
[INFO] [stdout]    --> src/tensor/mod.rs:686:8
[INFO] [stdout]     |
[INFO] [stdout] 560 | / impl<'a, T> Tensor<'a, T>
[INFO] [stdout] 561 | | where
[INFO] [stdout] 562 | |     T: TensorElement + Div<Output = T> + Sum<T>,
[INFO] [stdout] 563 | |     <T as FromStr>::Err: Error + 'static,
[INFO] [stdout] 564 | |     Vec<T>: IntoParallelIterator,
[INFO] [stdout] 565 | |     Vec<&'a T>: IntoParallelRefIterator<'a>,
[INFO] [stdout]     | |____________________________________________- methods in this implementation
[INFO] [stdout] ...
[INFO] [stdout] 686 |       fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec<T> {
[INFO] [stdout]     |          ^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout] 889 |       fn sum(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |          ^^^
[INFO] [stdout] ...
[INFO] [stdout] 904 |       fn prod(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |          ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `ReLU` should have a snake case name
[INFO] [stdout]   --> src/nn/activation.rs:25:8
[INFO] [stdout]    |
[INFO] [stdout] 25 | pub fn ReLU<'a, 'b, T>(x: &Tensor<'a, T>) -> Tensor<'b, T>
[INFO] [stdout]    |        ^^^^ help: convert the identifier to snake case: `re_lu`
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(non_snake_case)]` (part of `#[warn(nonstandard_style)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `PReLU` should have a snake case name
[INFO] [stdout]   --> src/nn/activation.rs:52:8
[INFO] [stdout]    |
[INFO] [stdout] 52 | pub fn PReLU<'a, 'b, T>(x: &Tensor<'a, T>, alpha: T) -> Tensor<'b, T>
[INFO] [stdout]    |        ^^^^^ help: convert the identifier to snake case: `pre_lu`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `Sigmoid` should have a snake case name
[INFO] [stdout]   --> src/nn/activation.rs:79:8
[INFO] [stdout]    |
[INFO] [stdout] 79 | pub fn Sigmoid<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case (notice the capitalization): `sigmoid`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `GeLU` should have a snake case name
[INFO] [stdout]    --> src/nn/activation.rs:108:8
[INFO] [stdout]     |
[INFO] [stdout] 108 | pub fn GeLU<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T>
[INFO] [stdout]     |        ^^^^ help: convert the identifier to snake case: `ge_lu`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: missing documentation for an associated function
[INFO] [stdout]   --> src/nn/loss.rs:14:5
[INFO] [stdout]    |
[INFO] [stdout] 14 |     fn loss(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T;
[INFO] [stdout]    |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout] note: the lint level is defined here
[INFO] [stdout]   --> src/nn/loss.rs:2:9
[INFO] [stdout]    |
[INFO] [stdout]  2 | #![warn(missing_docs)]
[INFO] [stdout]    |         ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: missing documentation for an associated function
[INFO] [stdout]   --> src/nn/loss.rs:15:5
[INFO] [stdout]    |
[INFO] [stdout] 15 |     fn backward();
[INFO] [stdout]    |     ^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `BCEntroypyLoss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:31:8
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^^^^^^^^^ help: convert the identifier to snake case: `bcentroypy_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `CEntroypyLoss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:50:8
[INFO] [stdout]    |
[INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^^^^^^^^ help: convert the identifier to snake case: `centroypy_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `L1Loss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:73:8
[INFO] [stdout]    |
[INFO] [stdout] 73 | pub fn L1Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^ help: convert the identifier to snake case: `l1_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `L2Loss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:96:8
[INFO] [stdout]    |
[INFO] [stdout] 96 | pub fn L2Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^ help: convert the identifier to snake case: `l2_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `MaxPool` should have a snake case name
[INFO] [stdout]   --> src/nn/pooling.rs:23:8
[INFO] [stdout]    |
[INFO] [stdout] 23 | pub fn MaxPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case: `max_pool`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `MinPool` should have a snake case name
[INFO] [stdout]   --> src/nn/pooling.rs:62:8
[INFO] [stdout]    |
[INFO] [stdout] 62 | pub fn MinPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case: `min_pool`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `AvgPool` should have a snake case name
[INFO] [stdout]   --> src/nn/pooling.rs:93:8
[INFO] [stdout]    |
[INFO] [stdout] 93 | pub fn AvgPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case: `avg_pool`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: missing documentation for a struct
[INFO] [stdout]  --> src/nn/transform.rs:4:1
[INFO] [stdout]   |
[INFO] [stdout] 4 | pub struct Transform;
[INFO] [stdout]   | ^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout] note: the lint level is defined here
[INFO] [stdout]  --> src/nn/transform.rs:2:9
[INFO] [stdout]   |
[INFO] [stdout] 2 | #![warn(missing_docs)]
[INFO] [stdout]   |         ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: missing documentation for a struct
[INFO] [stdout]    --> src/tensor/mod.rs:145:1
[INFO] [stdout]     |
[INFO] [stdout] 145 | pub struct Tensor<'a, T>
[INFO] [stdout]     | ^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] note: the lint level is defined here
[INFO] [stdout]    --> src/lib.rs:1:9
[INFO] [stdout]     |
[INFO] [stdout]   1 | #![warn(missing_docs)]
[INFO] [stdout]     |         ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `marker::PhantomData`
[INFO] [stdout]  --> src/nn/optimizer.rs:4:25
[INFO] [stdout]   |
[INFO] [stdout] 4 | use std::{error::Error, marker::PhantomData, str::FromStr};
[INFO] [stdout]   |                         ^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout]   = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `rayon::prelude::*`
[INFO] [stdout]  --> src/nn/optimizer.rs:7:5
[INFO] [stdout]   |
[INFO] [stdout] 7 | use rayon::prelude::*;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `idx`
[INFO] [stdout]   --> src/nn/dataset.rs:57:26
[INFO] [stdout]    |
[INFO] [stdout] 57 |     fn get<'a, T>(&self, idx: usize) -> (Tensor<'a, T>, String)
[INFO] [stdout]    |                          ^^^ help: if this is intentional, prefix it with an underscore: `_idx`
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(unused_variables)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y`
[INFO] [stdout]   --> src/nn/loss.rs:31:30
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                              ^ help: if this is intentional, prefix it with an underscore: `_y`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y_hat`
[INFO] [stdout]   --> src/nn/loss.rs:31:49
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                                                 ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y`
[INFO] [stdout]   --> src/nn/loss.rs:50:29
[INFO] [stdout]    |
[INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                             ^ help: if this is intentional, prefix it with an underscore: `_y`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y_hat`
[INFO] [stdout]   --> src/nn/loss.rs:50:48
[INFO] [stdout]    |
[INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                                                ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `cost`
[INFO] [stdout]   --> src/nn/optimizer.rs:75:31
[INFO] [stdout]    |
[INFO] [stdout] 75 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]    |                               ^^^^ help: if this is intentional, prefix it with an underscore: `_cost`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `vars`
[INFO] [stdout]   --> src/nn/optimizer.rs:75:40
[INFO] [stdout]    |
[INFO] [stdout] 75 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]    |                                        ^^^^ help: if this is intentional, prefix it with an underscore: `_vars`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `cost`
[INFO] [stdout]    --> src/nn/optimizer.rs:117:31
[INFO] [stdout]     |
[INFO] [stdout] 117 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]     |                               ^^^^ help: if this is intentional, prefix it with an underscore: `_cost`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `vars`
[INFO] [stdout]    --> src/nn/optimizer.rs:117:40
[INFO] [stdout]     |
[INFO] [stdout] 117 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]     |                                        ^^^^ help: if this is intentional, prefix it with an underscore: `_vars`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `tensor`
[INFO] [stdout]   --> src/nn/mod.rs:24:5
[INFO] [stdout]    |
[INFO] [stdout] 24 |     tensor: Tensor<'a, T>,
[INFO] [stdout]    |     ^^^^^^ help: if this is intentional, prefix it with an underscore: `_tensor`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `kernel`
[INFO] [stdout]   --> src/nn/mod.rs:25:5
[INFO] [stdout]    |
[INFO] [stdout] 25 |     kernel: Tensor<'a, T>,
[INFO] [stdout]    |     ^^^^^^ help: if this is intentional, prefix it with an underscore: `_kernel`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `stride`
[INFO] [stdout]   --> src/nn/mod.rs:26:5
[INFO] [stdout]    |
[INFO] [stdout] 26 |     stride: usize,
[INFO] [stdout]    |     ^^^^^^ help: if this is intentional, prefix it with an underscore: `_stride`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `padding`
[INFO] [stdout]   --> src/nn/mod.rs:27:5
[INFO] [stdout]    |
[INFO] [stdout] 27 |     padding: usize,
[INFO] [stdout]    |     ^^^^^^^ help: if this is intentional, prefix it with an underscore: `_padding`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `f`
[INFO] [stdout]    --> src/tensor/mod.rs:257:19
[INFO] [stdout]     |
[INFO] [stdout] 257 |     fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
[INFO] [stdout]     |                   ^ help: if this is intentional, prefix it with an underscore: `_f`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `start_idx`
[INFO] [stdout]    --> src/tensor/mod.rs:686:30
[INFO] [stdout]     |
[INFO] [stdout] 686 |     fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec<T> {
[INFO] [stdout]     |                              ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_start_idx`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `size`
[INFO] [stdout]    --> src/tensor/mod.rs:686:48
[INFO] [stdout]     |
[INFO] [stdout] 686 |     fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec<T> {
[INFO] [stdout]     |                                                ^^^^ help: if this is intentional, prefix it with an underscore: `_size`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `rowcol`
[INFO] [stdout]    --> src/tensor/mod.rs:889:19
[INFO] [stdout]     |
[INFO] [stdout] 889 |     fn sum(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                   ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `dimension`
[INFO] [stdout]    --> src/tensor/mod.rs:889:34
[INFO] [stdout]     |
[INFO] [stdout] 889 |     fn sum(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                                  ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `rowcol`
[INFO] [stdout]    --> src/tensor/mod.rs:904:20
[INFO] [stdout]     |
[INFO] [stdout] 904 |     fn prod(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                    ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `dimension`
[INFO] [stdout]    --> src/tensor/mod.rs:904:35
[INFO] [stdout]     |
[INFO] [stdout] 904 |     fn prod(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                                   ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: field `layers` is never read
[INFO] [stdout]   --> src/nn/mod.rs:39:5
[INFO] [stdout]    |
[INFO] [stdout] 37 | pub struct Net {
[INFO] [stdout]    |            --- field in this struct
[INFO] [stdout] 38 |     /// Represents all layers in the network
[INFO] [stdout] 39 |     layers: Vec<Layer>,
[INFO] [stdout]    |     ^^^^^^
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: fields `data`, `batch_size`, and `shuffle` are never read
[INFO] [stdout]   --> src/nn/dataloader.rs:10:5
[INFO] [stdout]    |
[INFO] [stdout]  8 | pub struct DataLoader<D: DataSet + Clone> {
[INFO] [stdout]    |            ---------- fields in this struct
[INFO] [stdout]  9 |     /// Dataset
[INFO] [stdout] 10 |     data: D,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 11 |     /// Batch size
[INFO] [stdout] 12 |     batch_size: usize,
[INFO] [stdout]    |     ^^^^^^^^^^
[INFO] [stdout] 13 |     /// Whether or not to shuffle data
[INFO] [stdout] 14 |     shuffle: bool,
[INFO] [stdout]    |     ^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout]    = note: `DataLoader` has a derived impl for the trait `Clone`, but this is intentionally ignored during dead code analysis
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated items `new` and `iter` are never used
[INFO] [stdout]   --> src/nn/dataloader.rs:41:8
[INFO] [stdout]    |
[INFO] [stdout] 33 | impl<D: DataSet + Clone> DataLoader<D> {
[INFO] [stdout]    | -------------------------------------- associated items in this implementation
[INFO] [stdout] ...
[INFO] [stdout] 41 |     fn new(data: D, batch_size: usize, shuffle: bool) -> DataLoader<D> {
[INFO] [stdout]    |        ^^^
[INFO] [stdout] ...
[INFO] [stdout] 56 |     fn iter(&mut self) -> DataLoader<D> {
[INFO] [stdout]    |        ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: fields `img_dir`, `transform`, and `target_transform` are never read
[INFO] [stdout]   --> src/nn/dataset.rs:33:5
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub struct CustomImageDataset {
[INFO] [stdout]    |            ------------------ fields in this struct
[INFO] [stdout] 32 |     img_labels: Vec<String>,
[INFO] [stdout] 33 |     img_dir: &'static str,
[INFO] [stdout]    |     ^^^^^^^
[INFO] [stdout] 34 |     transform: Option<Transform>,
[INFO] [stdout]    |     ^^^^^^^^^
[INFO] [stdout] 35 |     target_transform: Option<Transform>,
[INFO] [stdout]    |     ^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: multiple fields are never read
[INFO] [stdout]   --> src/nn/optimizer.rs:38:5
[INFO] [stdout]    |
[INFO] [stdout] 36 | pub struct Adam<T> {
[INFO] [stdout]    |            ---- fields in this struct
[INFO] [stdout] 37 |     /// Learning rate
[INFO] [stdout] 38 |     lr: T,
[INFO] [stdout]    |     ^^
[INFO] [stdout] 39 |     /// Momentum
[INFO] [stdout] 40 |     momentum: f32,
[INFO] [stdout]    |     ^^^^^^^^
[INFO] [stdout] 41 |     /// Decay rate
[INFO] [stdout] 42 |     decay_rate: f32,
[INFO] [stdout]    |     ^^^^^^^^^^
[INFO] [stdout] 43 |     /// Beta 1
[INFO] [stdout] 44 |     beta1: f32,
[INFO] [stdout]    |     ^^^^^
[INFO] [stdout] 45 |     /// Beta 2:
[INFO] [stdout] 46 |     beta2: f32,
[INFO] [stdout]    |     ^^^^^
[INFO] [stdout] 47 |     /// Epsilon
[INFO] [stdout] 48 |     epsilon: f32,
[INFO] [stdout]    |     ^^^^^^^
[INFO] [stdout] 49 |     m_dw: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 50 |     v_dw: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 51 |     m_db: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 52 |     v_db: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: fields `lr`, `momentum`, and `decay_rate` are never read
[INFO] [stdout]    --> src/nn/optimizer.rs:97:5
[INFO] [stdout]     |
[INFO] [stdout]  95 | pub struct SGD<T> {
[INFO] [stdout]     |            --- fields in this struct
[INFO] [stdout]  96 |     /// Learning rate
[INFO] [stdout]  97 |     lr: T,
[INFO] [stdout]     |     ^^
[INFO] [stdout]  98 |     /// Momentum
[INFO] [stdout]  99 |     momentum: f32,
[INFO] [stdout]     |     ^^^^^^^^
[INFO] [stdout] 100 |     /// Decay rate
[INFO] [stdout] 101 |     decay_rate: f32,
[INFO] [stdout]     |     ^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `swap` is never used
[INFO] [stdout]   --> src/tensor/mod.rs:33:4
[INFO] [stdout]    |
[INFO] [stdout] 33 | fn swap(lhs: &mut usize, rhs: &mut usize) {
[INFO] [stdout]    |    ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: methods `get_sub_tensor`, `sum`, and `prod` are never used
[INFO] [stdout]    --> src/tensor/mod.rs:686:8
[INFO] [stdout]     |
[INFO] [stdout] 560 | / impl<'a, T> Tensor<'a, T>
[INFO] [stdout] 561 | | where
[INFO] [stdout] 562 | |     T: TensorElement + Div<Output = T> + Sum<T>,
[INFO] [stdout] 563 | |     <T as FromStr>::Err: Error + 'static,
[INFO] [stdout] 564 | |     Vec<T>: IntoParallelIterator,
[INFO] [stdout] 565 | |     Vec<&'a T>: IntoParallelRefIterator<'a>,
[INFO] [stdout]     | |____________________________________________- methods in this implementation
[INFO] [stdout] ...
[INFO] [stdout] 686 |       fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec<T> {
[INFO] [stdout]     |          ^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout] 889 |       fn sum(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |          ^^^
[INFO] [stdout] ...
[INFO] [stdout] 904 |       fn prod(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |          ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `ReLU` should have a snake case name
[INFO] [stdout]   --> src/nn/activation.rs:25:8
[INFO] [stdout]    |
[INFO] [stdout] 25 | pub fn ReLU<'a, 'b, T>(x: &Tensor<'a, T>) -> Tensor<'b, T>
[INFO] [stdout]    |        ^^^^ help: convert the identifier to snake case: `re_lu`
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(non_snake_case)]` (part of `#[warn(nonstandard_style)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `PReLU` should have a snake case name
[INFO] [stdout]   --> src/nn/activation.rs:52:8
[INFO] [stdout]    |
[INFO] [stdout] 52 | pub fn PReLU<'a, 'b, T>(x: &Tensor<'a, T>, alpha: T) -> Tensor<'b, T>
[INFO] [stdout]    |        ^^^^^ help: convert the identifier to snake case: `pre_lu`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `Sigmoid` should have a snake case name
[INFO] [stdout]   --> src/nn/activation.rs:79:8
[INFO] [stdout]    |
[INFO] [stdout] 79 | pub fn Sigmoid<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case (notice the capitalization): `sigmoid`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `GeLU` should have a snake case name
[INFO] [stdout]    --> src/nn/activation.rs:108:8
[INFO] [stdout]     |
[INFO] [stdout] 108 | pub fn GeLU<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T>
[INFO] [stdout]     |        ^^^^ help: convert the identifier to snake case: `ge_lu`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: missing documentation for an associated function
[INFO] [stdout]   --> src/nn/loss.rs:14:5
[INFO] [stdout]    |
[INFO] [stdout] 14 |     fn loss(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T;
[INFO] [stdout]    |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout] note: the lint level is defined here
[INFO] [stdout]   --> src/nn/loss.rs:2:9
[INFO] [stdout]    |
[INFO] [stdout]  2 | #![warn(missing_docs)]
[INFO] [stdout]    |         ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: missing documentation for an associated function
[INFO] [stdout]   --> src/nn/loss.rs:15:5
[INFO] [stdout]    |
[INFO] [stdout] 15 |     fn backward();
[INFO] [stdout]    |     ^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `BCEntroypyLoss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:31:8
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^^^^^^^^^ help: convert the identifier to snake case: `bcentroypy_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `CEntroypyLoss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:50:8
[INFO] [stdout]    |
[INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^^^^^^^^ help: convert the identifier to snake case: `centroypy_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `L1Loss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:73:8
[INFO] [stdout]    |
[INFO] [stdout] 73 | pub fn L1Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^ help: convert the identifier to snake case: `l1_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `L2Loss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:96:8
[INFO] [stdout]    |
[INFO] [stdout] 96 | pub fn L2Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^ help: convert the identifier to snake case: `l2_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `MaxPool` should have a snake case name
[INFO] [stdout]   --> src/nn/pooling.rs:23:8
[INFO] [stdout]    |
[INFO] [stdout] 23 | pub fn MaxPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case: `max_pool`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `MinPool` should have a snake case name
[INFO] [stdout]   --> src/nn/pooling.rs:62:8
[INFO] [stdout]    |
[INFO] [stdout] 62 | pub fn MinPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case: `min_pool`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `AvgPool` should have a snake case name
[INFO] [stdout]   --> src/nn/pooling.rs:93:8
[INFO] [stdout]    |
[INFO] [stdout] 93 | pub fn AvgPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case: `avg_pool`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: missing documentation for a struct
[INFO] [stdout]  --> src/nn/transform.rs:4:1
[INFO] [stdout]   |
[INFO] [stdout] 4 | pub struct Transform;
[INFO] [stdout]   | ^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout] note: the lint level is defined here
[INFO] [stdout]  --> src/nn/transform.rs:2:9
[INFO] [stdout]   |
[INFO] [stdout] 2 | #![warn(missing_docs)]
[INFO] [stdout]   |         ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: missing documentation for a struct
[INFO] [stdout]    --> src/tensor/mod.rs:145:1
[INFO] [stdout]     |
[INFO] [stdout] 145 | pub struct Tensor<'a, T>
[INFO] [stdout]     | ^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] note: the lint level is defined here
[INFO] [stdout]    --> src/lib.rs:1:9
[INFO] [stdout]     |
[INFO] [stdout]   1 | #![warn(missing_docs)]
[INFO] [stdout]     |         ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stderr]     Finished `test` profile [unoptimized + debuginfo] target(s) in 24.00s
[INFO] running `Command { std: "docker" "inspect" "25623b942f78b7a8ad9c6331268b5e7e4728d9bb210f85be405d68f5c1a67c7f", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "25623b942f78b7a8ad9c6331268b5e7e4728d9bb210f85be405d68f5c1a67c7f", kill_on_drop: false }`
[INFO] [stdout] 25623b942f78b7a8ad9c6331268b5e7e4728d9bb210f85be405d68f5c1a67c7f
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:4848fb76d95f26979359cc7e45710b1dbc8f3acb7aeedee7c460d7702230f228" "/opt/rustwide/cargo-home/bin/cargo" "+c90bcb9571b7aab0d8beaa2ce8a998ffaf079d38" "test" "--frozen", kill_on_drop: false }`
[INFO] [stdout] eb535b81eebdc77ecd6723aaa84cff139be587891e79a12e0c9e6762985b15f0
[INFO] running `Command { std: "docker" "start" "-a" "eb535b81eebdc77ecd6723aaa84cff139be587891e79a12e0c9e6762985b15f0", kill_on_drop: false }`
[INFO] [stderr] warning: unused import: `marker::PhantomData`
[INFO] [stderr]  --> src/nn/optimizer.rs:4:25
[INFO] [stderr]   |
[INFO] [stderr] 4 | use std::{error::Error, marker::PhantomData, str::FromStr};
[INFO] [stderr]   |                         ^^^^^^^^^^^^^^^^^^^
[INFO] [stderr]   |
[INFO] [stderr]   = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default
[INFO] [stderr] 
[INFO] [stderr] warning: unused import: `rayon::prelude::*`
[INFO] [stderr]  --> src/nn/optimizer.rs:7:5
[INFO] [stderr]   |
[INFO] [stderr] 7 | use rayon::prelude::*;
[INFO] [stderr]   |     ^^^^^^^^^^^^^^^^^
[INFO] [stderr] 
[INFO] [stderr] warning: unused variable: `idx`
[INFO] [stderr]   --> src/nn/dataset.rs:57:26
[INFO] [stderr]    |
[INFO] [stderr] 57 |     fn get<'a, T>(&self, idx: usize) -> (Tensor<'a, T>, String)
[INFO] [stderr]    |                          ^^^ help: if this is intentional, prefix it with an underscore: `_idx`
[INFO] [stderr]    |
[INFO] [stderr]    = note: `#[warn(unused_variables)]` (part of `#[warn(unused)]`) on by default
[INFO] [stderr] 
[INFO] [stderr] warning: unused variable: `y`
[INFO] [stderr]   --> src/nn/loss.rs:31:30
[INFO] [stderr]    |
[INFO] [stderr] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stderr]    |                              ^ help: if this is intentional, prefix it with an underscore: `_y`
[INFO] [stderr] 
[INFO] [stderr] warning: unused variable: `y_hat`
[INFO] [stderr]   --> src/nn/loss.rs:31:49
[INFO] [stderr]    |
[INFO] [stderr] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stderr]    |                                                 ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat`
[INFO] [stderr] 
[INFO] [stderr] warning: unused variable: `y`
[INFO] [stderr]   --> src/nn/loss.rs:50:29
[INFO] [stderr]    |
[INFO] [stderr] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stderr]    |                             ^ help: if this is intentional, prefix it with an underscore: `_y`
[INFO] [stderr] 
[INFO] [stderr] warning: unused variable: `y_hat`
[INFO] [stderr]   --> src/nn/loss.rs:50:48
[INFO] [stderr]    |
[INFO] [stderr] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stderr]    |                                                ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat`
[INFO] [stderr] 
[INFO] [stderr] warning: unused variable: `cost`
[INFO] [stderr]   --> src/nn/optimizer.rs:75:31
[INFO] [stderr]    |
[INFO] [stderr] 75 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stderr]    |                               ^^^^ help: if this is intentional, prefix it with an underscore: `_cost`
[INFO] [stderr] 
[INFO] [stderr] warning: unused variable: `vars`
[INFO] [stderr]   --> src/nn/optimizer.rs:75:40
[INFO] [stderr]    |
[INFO] [stderr] 75 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stderr]    |                                        ^^^^ help: if this is intentional, prefix it with an underscore: `_vars`
[INFO] [stderr] 
[INFO] [stderr] warning: unused variable: `cost`
[INFO] [stderr]    --> src/nn/optimizer.rs:117:31
[INFO] [stderr]     |
[INFO] [stderr] 117 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stderr]     |                               ^^^^ help: if this is intentional, prefix it with an underscore: `_cost`
[INFO] [stderr] 
[INFO] [stderr] warning: unused variable: `vars`
[INFO] [stderr]    --> src/nn/optimizer.rs:117:40
[INFO] [stderr]     |
[INFO] [stderr] 117 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stderr]     |                                        ^^^^ help: if this is intentional, prefix it with an underscore: `_vars`
[INFO] [stderr] 
[INFO] [stderr] warning: unused variable: `tensor`
[INFO] [stderr]   --> src/nn/mod.rs:24:5
[INFO] [stderr]    |
[INFO] [stderr] 24 |     tensor: Tensor<'a, T>,
[INFO] [stderr]    |     ^^^^^^ help: if this is intentional, prefix it with an underscore: `_tensor`
[INFO] [stderr] 
[INFO] [stderr] warning: unused variable: `kernel`
[INFO] [stderr]   --> src/nn/mod.rs:25:5
[INFO] [stderr]    |
[INFO] [stderr] 25 |     kernel: Tensor<'a, T>,
[INFO] [stderr]    |     ^^^^^^ help: if this is intentional, prefix it with an underscore: `_kernel`
[INFO] [stderr] 
[INFO] [stderr] warning: unused variable: `stride`
[INFO] [stderr]   --> src/nn/mod.rs:26:5
[INFO] [stderr]    |
[INFO] [stderr] 26 |     stride: usize,
[INFO] [stderr]    |     ^^^^^^ help: if this is intentional, prefix it with an underscore: `_stride`
[INFO] [stderr] 
[INFO] [stderr] warning: unused variable: `padding`
[INFO] [stderr]   --> src/nn/mod.rs:27:5
[INFO] [stderr]    |
[INFO] [stderr] 27 |     padding: usize,
[INFO] [stderr]    |     ^^^^^^^ help: if this is intentional, prefix it with an underscore: `_padding`
[INFO] [stderr] 
[INFO] [stderr] warning: unused variable: `f`
[INFO] [stderr]    --> src/tensor/mod.rs:257:19
[INFO] [stderr]     |
[INFO] [stderr] 257 |     fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
[INFO] [stderr]     |                   ^ help: if this is intentional, prefix it with an underscore: `_f`
[INFO] [stderr] 
[INFO] [stderr] warning: unused variable: `start_idx`
[INFO] [stderr]    --> src/tensor/mod.rs:686:30
[INFO] [stderr]     |
[INFO] [stderr] 686 |     fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec<T> {
[INFO] [stderr]     |                              ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_start_idx`
[INFO] [stderr] 
[INFO] [stderr] warning: unused variable: `size`
[INFO] [stderr]    --> src/tensor/mod.rs:686:48
[INFO] [stderr]     |
[INFO] [stderr] 686 |     fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec<T> {
[INFO] [stderr]     |                                                ^^^^ help: if this is intentional, prefix it with an underscore: `_size`
[INFO] [stderr] 
[INFO] [stderr] warning: unused variable: `rowcol`
[INFO] [stderr]    --> src/tensor/mod.rs:889:19
[INFO] [stderr]     |
[INFO] [stderr] 889 |     fn sum(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stderr]     |                   ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol`
[INFO] [stderr] 
[INFO] [stderr] warning: unused variable: `dimension`
[INFO] [stderr]    --> src/tensor/mod.rs:889:34
[INFO] [stderr]     |
[INFO] [stderr] 889 |     fn sum(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stderr]     |                                  ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension`
[INFO] [stderr] 
[INFO] [stderr] warning: unused variable: `rowcol`
[INFO] [stderr]    --> src/tensor/mod.rs:904:20
[INFO] [stderr]     |
[INFO] [stderr] 904 |     fn prod(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stderr]     |                    ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol`
[INFO] [stderr] 
[INFO] [stderr] warning: unused variable: `dimension`
[INFO] [stderr]    --> src/tensor/mod.rs:904:35
[INFO] [stderr]     |
[INFO] [stderr] 904 |     fn prod(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stderr]     |                                   ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension`
[INFO] [stderr] 
[INFO] [stderr] warning: field `layers` is never read
[INFO] [stderr]   --> src/nn/mod.rs:39:5
[INFO] [stderr]    |
[INFO] [stderr] 37 | pub struct Net {
[INFO] [stderr]    |            --- field in this struct
[INFO] [stderr] 38 |     /// Represents all layers in the network
[INFO] [stderr] 39 |     layers: Vec<Layer>,
[INFO] [stderr]    |     ^^^^^^
[INFO] [stderr]    |
[INFO] [stderr]    = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default
[INFO] [stderr] 
[INFO] [stderr] warning: fields `data`, `batch_size`, and `shuffle` are never read
[INFO] [stderr]   --> src/nn/dataloader.rs:10:5
[INFO] [stderr]    |
[INFO] [stderr]  8 | pub struct DataLoader<D: DataSet + Clone> {
[INFO] [stderr]    |            ---------- fields in this struct
[INFO] [stderr]  9 |     /// Dataset
[INFO] [stderr] 10 |     data: D,
[INFO] [stderr]    |     ^^^^
[INFO] [stderr] 11 |     /// Batch size
[INFO] [stderr] 12 |     batch_size: usize,
[INFO] [stderr]    |     ^^^^^^^^^^
[INFO] [stderr] 13 |     /// Whether or not to shuffle data
[INFO] [stderr] 14 |     shuffle: bool,
[INFO] [stderr]    |     ^^^^^^^
[INFO] [stderr]    |
[INFO] [stderr]    = note: `DataLoader` has a derived impl for the trait `Clone`, but this is intentionally ignored during dead code analysis
[INFO] [stderr] 
[INFO] [stderr] warning: associated items `new` and `iter` are never used
[INFO] [stderr]   --> src/nn/dataloader.rs:41:8
[INFO] [stderr]    |
[INFO] [stderr] 33 | impl<D: DataSet + Clone> DataLoader<D> {
[INFO] [stderr]    | -------------------------------------- associated items in this implementation
[INFO] [stderr] ...
[INFO] [stderr] 41 |     fn new(data: D, batch_size: usize, shuffle: bool) -> DataLoader<D> {
[INFO] [stderr]    |        ^^^
[INFO] [stderr] ...
[INFO] [stderr] 56 |     fn iter(&mut self) -> DataLoader<D> {
[INFO] [stderr]    |        ^^^^
[INFO] [stderr] 
[INFO] [stderr] warning: fields `img_dir`, `transform`, and `target_transform` are never read
[INFO] [stderr]   --> src/nn/dataset.rs:33:5
[INFO] [stderr]    |
[INFO] [stderr] 31 | pub struct CustomImageDataset {
[INFO] [stderr]    |            ------------------ fields in this struct
[INFO] [stderr] 32 |     img_labels: Vec<String>,
[INFO] [stderr] 33 |     img_dir: &'static str,
[INFO] [stderr]    |     ^^^^^^^
[INFO] [stderr] 34 |     transform: Option<Transform>,
[INFO] [stderr]    |     ^^^^^^^^^
[INFO] [stderr] 35 |     target_transform: Option<Transform>,
[INFO] [stderr]    |     ^^^^^^^^^^^^^^^^
[INFO] [stderr] 
[INFO] [stderr] warning: multiple fields are never read
[INFO] [stderr]   --> src/nn/optimizer.rs:38:5
[INFO] [stderr]    |
[INFO] [stderr] 36 | pub struct Adam<T> {
[INFO] [stderr]    |            ---- fields in this struct
[INFO] [stderr] 37 |     /// Learning rate
[INFO] [stderr] 38 |     lr: T,
[INFO] [stderr]    |     ^^
[INFO] [stderr] 39 |     /// Momentum
[INFO] [stderr] 40 |     momentum: f32,
[INFO] [stderr]    |     ^^^^^^^^
[INFO] [stderr] 41 |     /// Decay rate
[INFO] [stderr] 42 |     decay_rate: f32,
[INFO] [stderr]    |     ^^^^^^^^^^
[INFO] [stderr] 43 |     /// Beta 1
[INFO] [stderr] 44 |     beta1: f32,
[INFO] [stderr]    |     ^^^^^
[INFO] [stderr] 45 |     /// Beta 2:
[INFO] [stderr] 46 |     beta2: f32,
[INFO] [stderr]    |     ^^^^^
[INFO] [stderr] 47 |     /// Epsilon
[INFO] [stderr] 48 |     epsilon: f32,
[INFO] [stderr]    |     ^^^^^^^
[INFO] [stderr] 49 |     m_dw: f32,
[INFO] [stderr]    |     ^^^^
[INFO] [stderr] 50 |     v_dw: f32,
[INFO] [stderr]    |     ^^^^
[INFO] [stderr] 51 |     m_db: f32,
[INFO] [stderr]    |     ^^^^
[INFO] [stderr] 52 |     v_db: f32,
[INFO] [stderr]    |     ^^^^
[INFO] [stderr] 
[INFO] [stderr] warning: fields `lr`, `momentum`, and `decay_rate` are never read
[INFO] [stderr]    --> src/nn/optimizer.rs:97:5
[INFO] [stderr]     |
[INFO] [stderr]  95 | pub struct SGD<T> {
[INFO] [stderr]     |            --- fields in this struct
[INFO] [stderr]  96 |     /// Learning rate
[INFO] [stderr]  97 |     lr: T,
[INFO] [stderr]     |     ^^
[INFO] [stderr]  98 |     /// Momentum
[INFO] [stderr]  99 |     momentum: f32,
[INFO] [stderr]     |     ^^^^^^^^
[INFO] [stderr] 100 |     /// Decay rate
[INFO] [stderr] 101 |     decay_rate: f32,
[INFO] [stderr]     |     ^^^^^^^^^^
[INFO] [stderr] 
[INFO] [stderr] warning: function `swap` is never used
[INFO] [stderr]   --> src/tensor/mod.rs:33:4
[INFO] [stderr]    |
[INFO] [stderr] 33 | fn swap(lhs: &mut usize, rhs: &mut usize) {
[INFO] [stderr]    |    ^^^^
[INFO] [stderr] 
[INFO] [stderr] warning: methods `get_sub_tensor`, `sum`, and `prod` are never used
[INFO] [stderr]    --> src/tensor/mod.rs:686:8
[INFO] [stderr]     |
[INFO] [stderr] 560 | / impl<'a, T> Tensor<'a, T>
[INFO] [stderr] 561 | | where
[INFO] [stderr] 562 | |     T: TensorElement + Div<Output = T> + Sum<T>,
[INFO] [stderr] 563 | |     <T as FromStr>::Err: Error + 'static,
[INFO] [stderr] 564 | |     Vec<T>: IntoParallelIterator,
[INFO] [stderr] 565 | |     Vec<&'a T>: IntoParallelRefIterator<'a>,
[INFO] [stderr]     | |____________________________________________- methods in this implementation
[INFO] [stderr] ...
[INFO] [stderr] 686 |       fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec<T> {
[INFO] [stderr]     |          ^^^^^^^^^^^^^^
[INFO] [stderr] ...
[INFO] [stderr] 889 |       fn sum(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stderr]     |          ^^^
[INFO] [stderr] ...
[INFO] [stderr] 904 |       fn prod(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stderr]     |          ^^^^
[INFO] [stderr] 
[INFO] [stderr] warning: function `ReLU` should have a snake case name
[INFO] [stderr]   --> src/nn/activation.rs:25:8
[INFO] [stderr]    |
[INFO] [stderr] 25 | pub fn ReLU<'a, 'b, T>(x: &Tensor<'a, T>) -> Tensor<'b, T>
[INFO] [stderr]    |        ^^^^ help: convert the identifier to snake case: `re_lu`
[INFO] [stderr]    |
[INFO] [stderr]    = note: `#[warn(non_snake_case)]` (part of `#[warn(nonstandard_style)]`) on by default
[INFO] [stderr] 
[INFO] [stderr] warning: function `PReLU` should have a snake case name
[INFO] [stderr]   --> src/nn/activation.rs:52:8
[INFO] [stderr]    |
[INFO] [stderr] 52 | pub fn PReLU<'a, 'b, T>(x: &Tensor<'a, T>, alpha: T) -> Tensor<'b, T>
[INFO] [stderr]    |        ^^^^^ help: convert the identifier to snake case: `pre_lu`
[INFO] [stderr] 
[INFO] [stderr] warning: function `Sigmoid` should have a snake case name
[INFO] [stderr]   --> src/nn/activation.rs:79:8
[INFO] [stderr]    |
[INFO] [stderr] 79 | pub fn Sigmoid<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T>
[INFO] [stderr]    |        ^^^^^^^ help: convert the identifier to snake case (notice the capitalization): `sigmoid`
[INFO] [stderr] 
[INFO] [stderr] warning: function `GeLU` should have a snake case name
[INFO] [stderr]    --> src/nn/activation.rs:108:8
[INFO] [stderr]     |
[INFO] [stderr] 108 | pub fn GeLU<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T>
[INFO] [stderr]     |        ^^^^ help: convert the identifier to snake case: `ge_lu`
[INFO] [stderr] 
[INFO] [stderr] warning: missing documentation for an associated function
[INFO] [stderr]   --> src/nn/loss.rs:14:5
[INFO] [stderr]    |
[INFO] [stderr] 14 |     fn loss(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T;
[INFO] [stderr]    |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stderr]    |
[INFO] [stderr] note: the lint level is defined here
[INFO] [stderr]   --> src/nn/loss.rs:2:9
[INFO] [stderr]    |
[INFO] [stderr]  2 | #![warn(missing_docs)]
[INFO] [stderr]    |         ^^^^^^^^^^^^
[INFO] [stderr] 
[INFO] [stderr] warning: missing documentation for an associated function
[INFO] [stderr]   --> src/nn/loss.rs:15:5
[INFO] [stderr]    |
[INFO] [stderr] 15 |     fn backward();
[INFO] [stderr]    |     ^^^^^^^^^^^^^^
[INFO] [stderr] 
[INFO] [stderr] warning: function `BCEntroypyLoss` should have a snake case name
[INFO] [stderr]   --> src/nn/loss.rs:31:8
[INFO] [stderr]    |
[INFO] [stderr] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stderr]    |        ^^^^^^^^^^^^^^ help: convert the identifier to snake case: `bcentroypy_loss`
[INFO] [stderr] 
[INFO] [stderr] warning: function `CEntroypyLoss` should have a snake case name
[INFO] [stderr]   --> src/nn/loss.rs:50:8
[INFO] [stderr]    |
[INFO] [stderr] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stderr]    |        ^^^^^^^^^^^^^ help: convert the identifier to snake case: `centroypy_loss`
[INFO] [stderr] 
[INFO] [stderr] warning: function `L1Loss` should have a snake case name
[INFO] [stderr]   --> src/nn/loss.rs:73:8
[INFO] [stderr]    |
[INFO] [stderr] 73 | pub fn L1Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stderr]    |        ^^^^^^ help: convert the identifier to snake case: `l1_loss`
[INFO] [stderr] 
[INFO] [stderr] warning: function `L2Loss` should have a snake case name
[INFO] [stderr]   --> src/nn/loss.rs:96:8
[INFO] [stderr]    |
[INFO] [stderr] 96 | pub fn L2Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stderr]    |        ^^^^^^ help: convert the identifier to snake case: `l2_loss`
[INFO] [stderr] 
[INFO] [stderr] warning: function `MaxPool` should have a snake case name
[INFO] [stderr]   --> src/nn/pooling.rs:23:8
[INFO] [stderr]    |
[INFO] [stderr] 23 | pub fn MaxPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T>
[INFO] [stderr]    |        ^^^^^^^ help: convert the identifier to snake case: `max_pool`
[INFO] [stderr] 
[INFO] [stderr] warning: function `MinPool` should have a snake case name
[INFO] [stderr]   --> src/nn/pooling.rs:62:8
[INFO] [stderr]    |
[INFO] [stderr] 62 | pub fn MinPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T>
[INFO] [stderr]    |        ^^^^^^^ help: convert the identifier to snake case: `min_pool`
[INFO] [stderr] 
[INFO] [stderr] warning: function `AvgPool` should have a snake case name
[INFO] [stderr]   --> src/nn/pooling.rs:93:8
[INFO] [stderr]    |
[INFO] [stderr] 93 | pub fn AvgPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T>
[INFO] [stderr]    |        ^^^^^^^ help: convert the identifier to snake case: `avg_pool`
[INFO] [stderr] 
[INFO] [stderr] warning: missing documentation for a struct
[INFO] [stderr]  --> src/nn/transform.rs:4:1
[INFO] [stderr]   |
[INFO] [stderr] 4 | pub struct Transform;
[INFO] [stderr]   | ^^^^^^^^^^^^^^^^^^^^
[INFO] [stderr]   |
[INFO] [stderr] note: the lint level is defined here
[INFO] [stderr]  --> src/nn/transform.rs:2:9
[INFO] [stderr]   |
[INFO] [stderr] 2 | #![warn(missing_docs)]
[INFO] [stderr]   |         ^^^^^^^^^^^^
[INFO] [stderr] 
[INFO] [stderr] warning: missing documentation for a struct
[INFO] [stderr]    --> src/tensor/mod.rs:145:1
[INFO] [stderr]     |
[INFO] [stderr] 145 | pub struct Tensor<'a, T>
[INFO] [stderr]     | ^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stderr]     |
[INFO] [stderr] note: the lint level is defined here
[INFO] [stderr]    --> src/lib.rs:1:9
[INFO] [stderr]     |
[INFO] [stderr]   1 | #![warn(missing_docs)]
[INFO] [stderr]     |         ^^^^^^^^^^^^
[INFO] [stderr] 
[INFO] [stderr] warning: `kaffe` (lib) generated 45 warnings (run `cargo fix --lib -p kaffe` to apply 22 suggestions)
[INFO] [stderr] warning: `kaffe` (lib test) generated 45 warnings (45 duplicates)
[INFO] [stderr]     Finished `test` profile [unoptimized + debuginfo] target(s) in 0.19s
[INFO] [stderr]      Running unittests src/lib.rs (/opt/rustwide/target/debug/deps/kaffe-97bede8cfcb8b023)
[INFO] [stdout] 
[INFO] [stdout] running 0 tests
[INFO] [stdout] 
[INFO] [stdout] test result: ok. 0 passed; 0 failed; 0 ignored; 0 measured; 0 filtered out; finished in 0.00s
[INFO] [stdout] 
[INFO] [stderr]      Running tests/nn.rs (/opt/rustwide/target/debug/deps/nn-1c88b5395785526c)
[INFO] [stdout] 
[INFO] [stdout] running 1 test
[INFO] [stdout] test relu ... ok
[INFO] [stdout] 
[INFO] [stdout] test result: ok. 1 passed; 0 failed; 0 ignored; 0 measured; 0 filtered out; finished in 0.00s
[INFO] [stdout] 
[INFO] [stderr]      Running tests/tensor.rs (/opt/rustwide/target/debug/deps/tensor-ccf94bfb46ee0ec6)
[INFO] [stdout] 
[INFO] [stderr]    Doc-tests kaffe
[INFO] [stdout] running 2 tests
[INFO] [stdout] test creation ... ok
[INFO] [stdout] test matmul ... ok
[INFO] [stdout] 
[INFO] [stdout] test result: ok. 2 passed; 0 failed; 0 ignored; 0 measured; 0 filtered out; finished in 0.00s
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] running 85 tests
[INFO] [stdout] test src/nn/dataloader.rs - nn::dataloader::DataLoader<D>::iter (line 53) ... ok
[INFO] [stdout] test src/nn/dataloader.rs - nn::dataloader::DataLoader<D>::next (line 25) ... ok
[INFO] [stdout] test src/nn/optimizer.rs - nn::optimizer::Adam (line 28) ... ok
[INFO] [stdout] test src/nn/dataloader.rs - nn::dataloader::DataLoader<D>::new (line 38) ... ok
[INFO] [stdout] test src/nn/loss.rs - nn::loss::CEntroypyLoss (line 43) ... ok
[INFO] [stdout] test src/nn/loss.rs - nn::loss::BCEntroypyLoss (line 24) ... ok
[INFO] [stdout] test src/nn/pooling.rs - nn::pooling::MaxPool (line 12) ... ok
[INFO] [stdout] test src/nn/optimizer.rs - nn::optimizer::SGD (line 87) ... ok
[INFO] [stdout] test src/nn/activation.rs - nn::activation::Sigmoid (line 71) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::abs (line 1245) ... ok
[INFO] [stdout] test src/nn/activation.rs - nn::activation::GeLU (line 100) ... ok
[INFO] [stdout] test src/nn/loss.rs - nn::loss::L1Loss (line 64) ... ok
[INFO] [stdout] test src/nn/loss.rs - nn::loss::L2Loss (line 87) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::add_self (line 1264) ... ok
[INFO] [stdout] test src/nn/activation.rs - nn::activation::ReLU (line 17) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::cosh (line 1209) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::abs_self (line 1348) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::avg (line 808) ... ok
[INFO] [stdout] test src/nn/pooling.rs - nn::pooling::AvgPool (line 81) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::add (line 922) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::all (line 1693) ... ok
[INFO] [stdout] test src/nn/pooling.rs - nn::pooling::MinPool (line 42) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::default (line 273) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::any (line 1675) ... ok
[INFO] [stdout] test src/nn/activation.rs - nn::activation::PReLU (line 44) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::div_val_self (line 1419) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::eigenvalue (line 1582) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::eye (line 341) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::div_val (line 1112) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::div (line 1030) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::count_where (line 1606) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::add_val_self (line 1365) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::cumprod (line 793) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::from_file (line 526) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::div_self (line 1327) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::add_val (line 1061) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::cumsum (line 778) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::from_slice (line 378) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::init (line 324) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::get_sub_tensor (line 681) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::find_all (line 1733) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::get (line 628) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::find (line 1711) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::get_vec_slice (line 656) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::identity (line 362) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::ln (line 1156) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::log (line 1136) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::max (line 737) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::matmul (line 1437) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::mean (line 829) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::min (line 757) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::median (line 844) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::mul_self (line 1306) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::mul (line 1003) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::ones (line 417) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::new (line 298) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::randomize_range (line 487) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::mul_val_self (line 1401) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::prod (line 897) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::ones_like (line 452) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::reshape (line 572) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::pow (line 1226) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::random_like (line 470) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::set (line 695) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::size (line 611) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::sub (line 949) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::sinh (line 1192) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::squeeze (line 593) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::set_where (line 1646) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::randomize (line 511) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::sub_abs (line 976) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::mm (line 1472) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::set_many (line 719) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::sum (line 883) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::mul_val (line 1095) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::transpose (line 1513) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::t (line 1548) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::sub_self (line 1285) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::sub_val (line 1078) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::zeros (line 400) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::transpose_copy (line 1564) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::tanh (line 1175) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::zeros_like (line 434) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::sub_val_self (line 1383) ... ok
[INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::sum_where (line 1624) ... ok
[INFO] [stdout] 
[INFO] [stdout] test result: ok. 85 passed; 0 failed; 0 ignored; 0 measured; 0 filtered out; finished in 6.61s
[INFO] [stdout] 
[INFO] running `Command { std: "docker" "inspect" "eb535b81eebdc77ecd6723aaa84cff139be587891e79a12e0c9e6762985b15f0", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "eb535b81eebdc77ecd6723aaa84cff139be587891e79a12e0c9e6762985b15f0", kill_on_drop: false }`
[INFO] [stdout] eb535b81eebdc77ecd6723aaa84cff139be587891e79a12e0c9e6762985b15f0
