[INFO] fetching crate kaffe 0.2.0...
[INFO] checking kaffe-0.2.0 against master#c2f2db79ca3024f68d22b45aa22b570775c2c4ad for pr-124157
[INFO] extracting crate kaffe 0.2.0 into /workspace/builds/worker-1-tc1/source
[INFO] validating manifest of crates.io crate kaffe 0.2.0 on toolchain c2f2db79ca3024f68d22b45aa22b570775c2c4ad
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+c2f2db79ca3024f68d22b45aa22b570775c2c4ad" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }`
[INFO] started tweaking crates.io crate kaffe 0.2.0
[INFO] finished tweaking crates.io crate kaffe 0.2.0
[INFO] tweaked toml for crates.io crate kaffe 0.2.0 written to /workspace/builds/worker-1-tc1/source/Cargo.toml
[INFO] crate crates.io crate kaffe 0.2.0 already has a lockfile, it will not be regenerated
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+c2f2db79ca3024f68d22b45aa22b570775c2c4ad" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-1-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-1-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:99613afd962a8cfa530ec1899472a458bd015a1ab0af876cf7eb06f6006d81ea" "/opt/rustwide/cargo-home/bin/cargo" "+c2f2db79ca3024f68d22b45aa22b570775c2c4ad" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }`
[INFO] [stdout] 560449195c541b378dc30bc4f7fdf50f0a1be07f15f7e214e4cba791e2fbff6c
[INFO] running `Command { std: "docker" "start" "-a" "560449195c541b378dc30bc4f7fdf50f0a1be07f15f7e214e4cba791e2fbff6c", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "inspect" "560449195c541b378dc30bc4f7fdf50f0a1be07f15f7e214e4cba791e2fbff6c", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "560449195c541b378dc30bc4f7fdf50f0a1be07f15f7e214e4cba791e2fbff6c", kill_on_drop: false }`
[INFO] [stdout] 560449195c541b378dc30bc4f7fdf50f0a1be07f15f7e214e4cba791e2fbff6c
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-1-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-1-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:99613afd962a8cfa530ec1899472a458bd015a1ab0af876cf7eb06f6006d81ea" "/opt/rustwide/cargo-home/bin/cargo" "+c2f2db79ca3024f68d22b45aa22b570775c2c4ad" "check" "--frozen" "--all" "--all-targets" "--message-format=json", kill_on_drop: false }`
[INFO] [stdout] 1cb98c66d491af650a3e8a835809619c3840a0a23092f05fa8af21c5e1f37292
[INFO] running `Command { std: "docker" "start" "-a" "1cb98c66d491af650a3e8a835809619c3840a0a23092f05fa8af21c5e1f37292", kill_on_drop: false }`
[INFO] [stderr]    Compiling libc v0.2.146
[INFO] [stderr]    Compiling proc-macro2 v1.0.60
[INFO] [stderr]    Compiling quote v1.0.28
[INFO] [stderr]    Compiling unicode-ident v1.0.9
[INFO] [stderr]    Compiling crossbeam-utils v0.8.15
[INFO] [stderr]    Compiling serde v1.0.164
[INFO] [stderr]    Compiling crossbeam-epoch v0.9.14
[INFO] [stderr]    Compiling num-traits v0.2.15
[INFO] [stderr]    Compiling rayon-core v1.11.0
[INFO] [stderr]     Checking either v1.8.1
[INFO] [stderr]    Compiling rustix v0.37.19
[INFO] [stderr]    Compiling serde_json v1.0.96
[INFO] [stderr]     Checking itertools v0.10.5
[INFO] [stderr]    Compiling anyhow v1.0.71
[INFO] [stderr]     Checking plotters-backend v0.3.4
[INFO] [stderr]     Checking crossbeam-channel v0.5.8
[INFO] [stderr]     Checking ciborium-io v0.2.1
[INFO] [stderr]     Checking crossbeam-deque v0.8.3
[INFO] [stderr]     Checking clap_lex v0.5.0
[INFO] [stderr]    Compiling syn v2.0.18
[INFO] [stderr]     Checking half v1.8.2
[INFO] [stderr]     Checking itoa v1.0.6
[INFO] [stderr]     Checking ryu v1.0.13
[INFO] [stderr]     Checking anstyle v1.0.0
[INFO] [stderr]     Checking linux-raw-sys v0.3.8
[INFO] [stderr]     Checking ciborium-ll v0.2.1
[INFO] [stderr]     Checking plotters-svg v0.3.3
[INFO] [stderr]     Checking clap_builder v4.3.3
[INFO] [stderr]     Checking regex-syntax v0.7.2
[INFO] [stderr]     Checking walkdir v2.3.3
[INFO] [stderr]     Checking plotters v0.3.4
[INFO] [stderr]     Checking criterion-plot v0.5.0
[INFO] [stderr]     Checking once_cell v1.18.0
[INFO] [stderr]     Checking num_cpus v1.15.0
[INFO] [stderr]     Checking getrandom v0.2.10
[INFO] [stderr]     Checking io-lifetimes v1.0.11
[INFO] [stderr]     Checking rand_core v0.6.4
[INFO] [stderr]     Checking rand_chacha v0.3.1
[INFO] [stderr]     Checking rand v0.8.5
[INFO] [stderr]     Checking rayon v1.7.0
[INFO] [stderr]     Checking regex v1.8.4
[INFO] [stderr]     Checking clap v4.3.3
[INFO] [stderr]    Compiling serde_derive v1.0.164
[INFO] [stderr]     Checking is-terminal v0.4.7
[INFO] [stderr]     Checking ciborium v0.2.1
[INFO] [stderr]     Checking kaffe v0.2.0 (/opt/rustwide/workdir)
[INFO] [stdout] warning: unused import: `marker::PhantomData`
[INFO] [stdout]  --> src/nn/optimizer.rs:4:25
[INFO] [stdout]   |
[INFO] [stdout] 4 | use std::{error::Error, marker::PhantomData, str::FromStr};
[INFO] [stdout]   |                         ^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout]   = note: `#[warn(unused_imports)]` on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `rayon::prelude::*`
[INFO] [stdout]  --> src/nn/optimizer.rs:7:5
[INFO] [stdout]   |
[INFO] [stdout] 7 | use rayon::prelude::*;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `idx`
[INFO] [stdout]   --> src/nn/dataset.rs:57:26
[INFO] [stdout]    |
[INFO] [stdout] 57 |     fn get<'a, T>(&self, idx: usize) -> (Tensor<'a, T>, String)
[INFO] [stdout]    |                          ^^^ help: if this is intentional, prefix it with an underscore: `_idx`
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(unused_variables)]` on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y`
[INFO] [stdout]   --> src/nn/loss.rs:31:30
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                              ^ help: if this is intentional, prefix it with an underscore: `_y`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y_hat`
[INFO] [stdout]   --> src/nn/loss.rs:31:49
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                                                 ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y`
[INFO] [stdout]   --> src/nn/loss.rs:50:29
[INFO] [stdout]    |
[INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                             ^ help: if this is intentional, prefix it with an underscore: `_y`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y_hat`
[INFO] [stdout]   --> src/nn/loss.rs:50:48
[INFO] [stdout]    |
[INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                                                ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `cost`
[INFO] [stdout]   --> src/nn/optimizer.rs:75:31
[INFO] [stdout]    |
[INFO] [stdout] 75 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]    |                               ^^^^ help: if this is intentional, prefix it with an underscore: `_cost`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `vars`
[INFO] [stdout]   --> src/nn/optimizer.rs:75:40
[INFO] [stdout]    |
[INFO] [stdout] 75 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]    |                                        ^^^^ help: if this is intentional, prefix it with an underscore: `_vars`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `cost`
[INFO] [stdout]    --> src/nn/optimizer.rs:117:31
[INFO] [stdout]     |
[INFO] [stdout] 117 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]     |                               ^^^^ help: if this is intentional, prefix it with an underscore: `_cost`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `vars`
[INFO] [stdout]    --> src/nn/optimizer.rs:117:40
[INFO] [stdout]     |
[INFO] [stdout] 117 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]     |                                        ^^^^ help: if this is intentional, prefix it with an underscore: `_vars`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `tensor`
[INFO] [stdout]   --> src/nn/mod.rs:24:5
[INFO] [stdout]    |
[INFO] [stdout] 24 |     tensor: Tensor<'a, T>,
[INFO] [stdout]    |     ^^^^^^ help: if this is intentional, prefix it with an underscore: `_tensor`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `kernel`
[INFO] [stdout]   --> src/nn/mod.rs:25:5
[INFO] [stdout]    |
[INFO] [stdout] 25 |     kernel: Tensor<'a, T>,
[INFO] [stdout]    |     ^^^^^^ help: if this is intentional, prefix it with an underscore: `_kernel`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `stride`
[INFO] [stdout]   --> src/nn/mod.rs:26:5
[INFO] [stdout]    |
[INFO] [stdout] 26 |     stride: usize,
[INFO] [stdout]    |     ^^^^^^ help: if this is intentional, prefix it with an underscore: `_stride`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `padding`
[INFO] [stdout]   --> src/nn/mod.rs:27:5
[INFO] [stdout]    |
[INFO] [stdout] 27 |     padding: usize,
[INFO] [stdout]    |     ^^^^^^^ help: if this is intentional, prefix it with an underscore: `_padding`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `f`
[INFO] [stdout]    --> src/tensor/mod.rs:257:19
[INFO] [stdout]     |
[INFO] [stdout] 257 |     fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
[INFO] [stdout]     |                   ^ help: if this is intentional, prefix it with an underscore: `_f`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `start_idx`
[INFO] [stdout]    --> src/tensor/mod.rs:686:30
[INFO] [stdout]     |
[INFO] [stdout] 686 |     fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec<T> {
[INFO] [stdout]     |                              ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_start_idx`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `size`
[INFO] [stdout]    --> src/tensor/mod.rs:686:48
[INFO] [stdout]     |
[INFO] [stdout] 686 |     fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec<T> {
[INFO] [stdout]     |                                                ^^^^ help: if this is intentional, prefix it with an underscore: `_size`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `rowcol`
[INFO] [stdout]    --> src/tensor/mod.rs:889:19
[INFO] [stdout]     |
[INFO] [stdout] 889 |     fn sum(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                   ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `dimension`
[INFO] [stdout]    --> src/tensor/mod.rs:889:34
[INFO] [stdout]     |
[INFO] [stdout] 889 |     fn sum(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                                  ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `rowcol`
[INFO] [stdout]    --> src/tensor/mod.rs:904:20
[INFO] [stdout]     |
[INFO] [stdout] 904 |     fn prod(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                    ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `dimension`
[INFO] [stdout]    --> src/tensor/mod.rs:904:35
[INFO] [stdout]     |
[INFO] [stdout] 904 |     fn prod(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                                   ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: field `layers` is never read
[INFO] [stdout]   --> src/nn/mod.rs:39:5
[INFO] [stdout]    |
[INFO] [stdout] 37 | pub struct Net {
[INFO] [stdout]    |            --- field in this struct
[INFO] [stdout] 38 |     /// Represents all layers in the network
[INFO] [stdout] 39 |     layers: Vec<Layer>,
[INFO] [stdout]    |     ^^^^^^
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(dead_code)]` on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: fields `data`, `batch_size`, and `shuffle` are never read
[INFO] [stdout]   --> src/nn/dataloader.rs:10:5
[INFO] [stdout]    |
[INFO] [stdout] 8  | pub struct DataLoader<D: DataSet + Clone> {
[INFO] [stdout]    |            ---------- fields in this struct
[INFO] [stdout] 9  |     /// Dataset
[INFO] [stdout] 10 |     data: D,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 11 |     /// Batch size
[INFO] [stdout] 12 |     batch_size: usize,
[INFO] [stdout]    |     ^^^^^^^^^^
[INFO] [stdout] 13 |     /// Whether or not to shuffle data
[INFO] [stdout] 14 |     shuffle: bool,
[INFO] [stdout]    |     ^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout]    = note: `DataLoader` has a derived impl for the trait `Clone`, but this is intentionally ignored during dead code analysis
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated items `new` and `iter` are never used
[INFO] [stdout]   --> src/nn/dataloader.rs:41:8
[INFO] [stdout]    |
[INFO] [stdout] 33 | impl<D: DataSet + Clone> DataLoader<D> {
[INFO] [stdout]    | -------------------------------------- associated items in this implementation
[INFO] [stdout] ...
[INFO] [stdout] 41 |     fn new(data: D, batch_size: usize, shuffle: bool) -> DataLoader<D> {
[INFO] [stdout]    |        ^^^
[INFO] [stdout] ...
[INFO] [stdout] 56 |     fn iter(&mut self) -> DataLoader<D> {
[INFO] [stdout]    |        ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: fields `img_dir`, `transform`, and `target_transform` are never read
[INFO] [stdout]   --> src/nn/dataset.rs:33:5
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub struct CustomImageDataset {
[INFO] [stdout]    |            ------------------ fields in this struct
[INFO] [stdout] 32 |     img_labels: Vec<String>,
[INFO] [stdout] 33 |     img_dir: &'static str,
[INFO] [stdout]    |     ^^^^^^^
[INFO] [stdout] 34 |     transform: Option<Transform>,
[INFO] [stdout]    |     ^^^^^^^^^
[INFO] [stdout] 35 |     target_transform: Option<Transform>,
[INFO] [stdout]    |     ^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: multiple fields are never read
[INFO] [stdout]   --> src/nn/optimizer.rs:38:5
[INFO] [stdout]    |
[INFO] [stdout] 36 | pub struct Adam<T> {
[INFO] [stdout]    |            ---- fields in this struct
[INFO] [stdout] 37 |     /// Learning rate
[INFO] [stdout] 38 |     lr: T,
[INFO] [stdout]    |     ^^
[INFO] [stdout] 39 |     /// Momentum
[INFO] [stdout] 40 |     momentum: f32,
[INFO] [stdout]    |     ^^^^^^^^
[INFO] [stdout] 41 |     /// Decay rate
[INFO] [stdout] 42 |     decay_rate: f32,
[INFO] [stdout]    |     ^^^^^^^^^^
[INFO] [stdout] 43 |     /// Beta 1
[INFO] [stdout] 44 |     beta1: f32,
[INFO] [stdout]    |     ^^^^^
[INFO] [stdout] 45 |     /// Beta 2:
[INFO] [stdout] 46 |     beta2: f32,
[INFO] [stdout]    |     ^^^^^
[INFO] [stdout] 47 |     /// Epsilon
[INFO] [stdout] 48 |     epsilon: f32,
[INFO] [stdout]    |     ^^^^^^^
[INFO] [stdout] 49 |     m_dw: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 50 |     v_dw: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 51 |     m_db: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 52 |     v_db: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: fields `lr`, `momentum`, and `decay_rate` are never read
[INFO] [stdout]    --> src/nn/optimizer.rs:97:5
[INFO] [stdout]     |
[INFO] [stdout] 95  | pub struct SGD<T> {
[INFO] [stdout]     |            --- fields in this struct
[INFO] [stdout] 96  |     /// Learning rate
[INFO] [stdout] 97  |     lr: T,
[INFO] [stdout]     |     ^^
[INFO] [stdout] 98  |     /// Momentum
[INFO] [stdout] 99  |     momentum: f32,
[INFO] [stdout]     |     ^^^^^^^^
[INFO] [stdout] 100 |     /// Decay rate
[INFO] [stdout] 101 |     decay_rate: f32,
[INFO] [stdout]     |     ^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `swap` is never used
[INFO] [stdout]   --> src/tensor/mod.rs:33:4
[INFO] [stdout]    |
[INFO] [stdout] 33 | fn swap(lhs: &mut usize, rhs: &mut usize) {
[INFO] [stdout]    |    ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: methods `get_sub_tensor`, `sum`, and `prod` are never used
[INFO] [stdout]    --> src/tensor/mod.rs:686:8
[INFO] [stdout]     |
[INFO] [stdout] 560 | / impl<'a, T> Tensor<'a, T>
[INFO] [stdout] 561 | | where
[INFO] [stdout] 562 | |     T: TensorElement + Div<Output = T> + Sum<T>,
[INFO] [stdout] 563 | |     <T as FromStr>::Err: Error + 'static,
[INFO] [stdout] 564 | |     Vec<T>: IntoParallelIterator,
[INFO] [stdout] 565 | |     Vec<&'a T>: IntoParallelRefIterator<'a>,
[INFO] [stdout]     | |____________________________________________- methods in this implementation
[INFO] [stdout] ...
[INFO] [stdout] 686 |       fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec<T> {
[INFO] [stdout]     |          ^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout] 889 |       fn sum(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |          ^^^
[INFO] [stdout] ...
[INFO] [stdout] 904 |       fn prod(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |          ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `ReLU` should have a snake case name
[INFO] [stdout]   --> src/nn/activation.rs:25:8
[INFO] [stdout]    |
[INFO] [stdout] 25 | pub fn ReLU<'a, 'b, T>(x: &Tensor<'a, T>) -> Tensor<'b, T>
[INFO] [stdout]    |        ^^^^ help: convert the identifier to snake case: `re_lu`
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(non_snake_case)]` on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `PReLU` should have a snake case name
[INFO] [stdout]   --> src/nn/activation.rs:52:8
[INFO] [stdout]    |
[INFO] [stdout] 52 | pub fn PReLU<'a, 'b, T>(x: &Tensor<'a, T>, alpha: T) -> Tensor<'b, T>
[INFO] [stdout]    |        ^^^^^ help: convert the identifier to snake case: `pre_lu`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `Sigmoid` should have a snake case name
[INFO] [stdout]   --> src/nn/activation.rs:79:8
[INFO] [stdout]    |
[INFO] [stdout] 79 | pub fn Sigmoid<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case (notice the capitalization): `sigmoid`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `GeLU` should have a snake case name
[INFO] [stdout]    --> src/nn/activation.rs:108:8
[INFO] [stdout]     |
[INFO] [stdout] 108 | pub fn GeLU<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T>
[INFO] [stdout]     |        ^^^^ help: convert the identifier to snake case: `ge_lu`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: missing documentation for an associated function
[INFO] [stdout]   --> src/nn/loss.rs:14:5
[INFO] [stdout]    |
[INFO] [stdout] 14 |     fn loss(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T;
[INFO] [stdout]    |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout] note: the lint level is defined here
[INFO] [stdout]   --> src/nn/loss.rs:2:9
[INFO] [stdout]    |
[INFO] [stdout] 2  | #![warn(missing_docs)]
[INFO] [stdout]    |         ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: missing documentation for an associated function
[INFO] [stdout]   --> src/nn/loss.rs:15:5
[INFO] [stdout]    |
[INFO] [stdout] 15 |     fn backward();
[INFO] [stdout]    |     ^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `BCEntroypyLoss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:31:8
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^^^^^^^^^ help: convert the identifier to snake case: `bcentroypy_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `CEntroypyLoss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:50:8
[INFO] [stdout]    |
[INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^^^^^^^^ help: convert the identifier to snake case: `centroypy_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `L1Loss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:73:8
[INFO] [stdout]    |
[INFO] [stdout] 73 | pub fn L1Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^ help: convert the identifier to snake case: `l1_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `L2Loss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:96:8
[INFO] [stdout]    |
[INFO] [stdout] 96 | pub fn L2Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^ help: convert the identifier to snake case: `l2_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `MaxPool` should have a snake case name
[INFO] [stdout]   --> src/nn/pooling.rs:23:8
[INFO] [stdout]    |
[INFO] [stdout] 23 | pub fn MaxPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case: `max_pool`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `MinPool` should have a snake case name
[INFO] [stdout]   --> src/nn/pooling.rs:62:8
[INFO] [stdout]    |
[INFO] [stdout] 62 | pub fn MinPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case: `min_pool`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `AvgPool` should have a snake case name
[INFO] [stdout]   --> src/nn/pooling.rs:93:8
[INFO] [stdout]    |
[INFO] [stdout] 93 | pub fn AvgPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case: `avg_pool`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: missing documentation for a struct
[INFO] [stdout]  --> src/nn/transform.rs:4:1
[INFO] [stdout]   |
[INFO] [stdout] 4 | pub struct Transform;
[INFO] [stdout]   | ^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout] note: the lint level is defined here
[INFO] [stdout]  --> src/nn/transform.rs:2:9
[INFO] [stdout]   |
[INFO] [stdout] 2 | #![warn(missing_docs)]
[INFO] [stdout]   |         ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: missing documentation for a struct
[INFO] [stdout]    --> src/tensor/mod.rs:145:1
[INFO] [stdout]     |
[INFO] [stdout] 145 | pub struct Tensor<'a, T>
[INFO] [stdout]     | ^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] note: the lint level is defined here
[INFO] [stdout]    --> src/lib.rs:1:9
[INFO] [stdout]     |
[INFO] [stdout] 1   | #![warn(missing_docs)]
[INFO] [stdout]     |         ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: 45 warnings emitted
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stderr]     Checking tinytemplate v1.2.1
[INFO] [stderr]     Checking criterion v0.5.1
[INFO] [stdout] error[E0432]: unresolved imports `kaffe::Matrix`, `kaffe::MatrixLinAlg`
[INFO] [stdout]  --> benches/matmul_benchmark.rs:1:13
[INFO] [stdout]   |
[INFO] [stdout] 1 | use kaffe::{Matrix, MatrixLinAlg};
[INFO] [stdout]   |             ^^^^^^  ^^^^^^^^^^^^ no `MatrixLinAlg` in the root
[INFO] [stdout]   |             |
[INFO] [stdout]   |             no `Matrix` in the root
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `marker::PhantomData`
[INFO] [stdout]  --> src/nn/optimizer.rs:4:25
[INFO] [stdout]   |
[INFO] [stdout] 4 | use std::{error::Error, marker::PhantomData, str::FromStr};
[INFO] [stdout]   |                         ^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout]   = note: `#[warn(unused_imports)]` on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `rayon::prelude::*`
[INFO] [stdout]  --> src/nn/optimizer.rs:7:5
[INFO] [stdout]   |
[INFO] [stdout] 7 | use rayon::prelude::*;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: aborting due to 1 previous error
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] For more information about this error, try `rustc --explain E0432`.
[INFO] [stdout] 
[INFO] [stderr] error: could not compile `kaffe` (bench "matmul_benchmark") due to 2 previous errors
[INFO] [stderr] warning: build failed, waiting for other jobs to finish...
[INFO] [stdout] error[E0432]: unresolved imports `kaffe::Matrix`, `kaffe::MatrixLinAlg`
[INFO] [stdout]  --> benches/matrix_ops_benches.rs:1:13
[INFO] [stdout]   |
[INFO] [stdout] 1 | use kaffe::{Matrix, MatrixLinAlg};
[INFO] [stdout]   |             ^^^^^^  ^^^^^^^^^^^^ no `MatrixLinAlg` in the root
[INFO] [stdout]   |             |
[INFO] [stdout]   |             no `Matrix` in the root
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: aborting due to 1 previous error
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] For more information about this error, try `rustc --explain E0432`.
[INFO] [stdout] 
[INFO] [stderr] error: could not compile `kaffe` (bench "matrix_ops_benches") due to 2 previous errors
[INFO] [stdout] warning: unused variable: `idx`
[INFO] [stdout]   --> src/nn/dataset.rs:57:26
[INFO] [stdout]    |
[INFO] [stdout] 57 |     fn get<'a, T>(&self, idx: usize) -> (Tensor<'a, T>, String)
[INFO] [stdout]    |                          ^^^ help: if this is intentional, prefix it with an underscore: `_idx`
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(unused_variables)]` on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y`
[INFO] [stdout]   --> src/nn/loss.rs:31:30
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                              ^ help: if this is intentional, prefix it with an underscore: `_y`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y_hat`
[INFO] [stdout]   --> src/nn/loss.rs:31:49
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                                                 ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y`
[INFO] [stdout]   --> src/nn/loss.rs:50:29
[INFO] [stdout]    |
[INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                             ^ help: if this is intentional, prefix it with an underscore: `_y`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y_hat`
[INFO] [stdout]   --> src/nn/loss.rs:50:48
[INFO] [stdout]    |
[INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                                                ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `cost`
[INFO] [stdout]   --> src/nn/optimizer.rs:75:31
[INFO] [stdout]    |
[INFO] [stdout] 75 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]    |                               ^^^^ help: if this is intentional, prefix it with an underscore: `_cost`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `vars`
[INFO] [stdout]   --> src/nn/optimizer.rs:75:40
[INFO] [stdout]    |
[INFO] [stdout] 75 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]    |                                        ^^^^ help: if this is intentional, prefix it with an underscore: `_vars`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `cost`
[INFO] [stdout]    --> src/nn/optimizer.rs:117:31
[INFO] [stdout]     |
[INFO] [stdout] 117 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]     |                               ^^^^ help: if this is intentional, prefix it with an underscore: `_cost`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `vars`
[INFO] [stdout]    --> src/nn/optimizer.rs:117:40
[INFO] [stdout]     |
[INFO] [stdout] 117 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]     |                                        ^^^^ help: if this is intentional, prefix it with an underscore: `_vars`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `tensor`
[INFO] [stdout]   --> src/nn/mod.rs:24:5
[INFO] [stdout]    |
[INFO] [stdout] 24 |     tensor: Tensor<'a, T>,
[INFO] [stdout]    |     ^^^^^^ help: if this is intentional, prefix it with an underscore: `_tensor`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `kernel`
[INFO] [stdout]   --> src/nn/mod.rs:25:5
[INFO] [stdout]    |
[INFO] [stdout] 25 |     kernel: Tensor<'a, T>,
[INFO] [stdout]    |     ^^^^^^ help: if this is intentional, prefix it with an underscore: `_kernel`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `stride`
[INFO] [stdout]   --> src/nn/mod.rs:26:5
[INFO] [stdout]    |
[INFO] [stdout] 26 |     stride: usize,
[INFO] [stdout]    |     ^^^^^^ help: if this is intentional, prefix it with an underscore: `_stride`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `padding`
[INFO] [stdout]   --> src/nn/mod.rs:27:5
[INFO] [stdout]    |
[INFO] [stdout] 27 |     padding: usize,
[INFO] [stdout]    |     ^^^^^^^ help: if this is intentional, prefix it with an underscore: `_padding`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `f`
[INFO] [stdout]    --> src/tensor/mod.rs:257:19
[INFO] [stdout]     |
[INFO] [stdout] 257 |     fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
[INFO] [stdout]     |                   ^ help: if this is intentional, prefix it with an underscore: `_f`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `start_idx`
[INFO] [stdout]    --> src/tensor/mod.rs:686:30
[INFO] [stdout]     |
[INFO] [stdout] 686 |     fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec<T> {
[INFO] [stdout]     |                              ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_start_idx`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `size`
[INFO] [stdout]    --> src/tensor/mod.rs:686:48
[INFO] [stdout]     |
[INFO] [stdout] 686 |     fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec<T> {
[INFO] [stdout]     |                                                ^^^^ help: if this is intentional, prefix it with an underscore: `_size`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `rowcol`
[INFO] [stdout]    --> src/tensor/mod.rs:889:19
[INFO] [stdout]     |
[INFO] [stdout] 889 |     fn sum(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                   ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `dimension`
[INFO] [stdout]    --> src/tensor/mod.rs:889:34
[INFO] [stdout]     |
[INFO] [stdout] 889 |     fn sum(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                                  ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `rowcol`
[INFO] [stdout]    --> src/tensor/mod.rs:904:20
[INFO] [stdout]     |
[INFO] [stdout] 904 |     fn prod(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                    ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `dimension`
[INFO] [stdout]    --> src/tensor/mod.rs:904:35
[INFO] [stdout]     |
[INFO] [stdout] 904 |     fn prod(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                                   ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: field `layers` is never read
[INFO] [stdout]   --> src/nn/mod.rs:39:5
[INFO] [stdout]    |
[INFO] [stdout] 37 | pub struct Net {
[INFO] [stdout]    |            --- field in this struct
[INFO] [stdout] 38 |     /// Represents all layers in the network
[INFO] [stdout] 39 |     layers: Vec<Layer>,
[INFO] [stdout]    |     ^^^^^^
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(dead_code)]` on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: fields `data`, `batch_size`, and `shuffle` are never read
[INFO] [stdout]   --> src/nn/dataloader.rs:10:5
[INFO] [stdout]    |
[INFO] [stdout] 8  | pub struct DataLoader<D: DataSet + Clone> {
[INFO] [stdout]    |            ---------- fields in this struct
[INFO] [stdout] 9  |     /// Dataset
[INFO] [stdout] 10 |     data: D,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 11 |     /// Batch size
[INFO] [stdout] 12 |     batch_size: usize,
[INFO] [stdout]    |     ^^^^^^^^^^
[INFO] [stdout] 13 |     /// Whether or not to shuffle data
[INFO] [stdout] 14 |     shuffle: bool,
[INFO] [stdout]    |     ^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout]    = note: `DataLoader` has a derived impl for the trait `Clone`, but this is intentionally ignored during dead code analysis
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated items `new` and `iter` are never used
[INFO] [stdout]   --> src/nn/dataloader.rs:41:8
[INFO] [stdout]    |
[INFO] [stdout] 33 | impl<D: DataSet + Clone> DataLoader<D> {
[INFO] [stdout]    | -------------------------------------- associated items in this implementation
[INFO] [stdout] ...
[INFO] [stdout] 41 |     fn new(data: D, batch_size: usize, shuffle: bool) -> DataLoader<D> {
[INFO] [stdout]    |        ^^^
[INFO] [stdout] ...
[INFO] [stdout] 56 |     fn iter(&mut self) -> DataLoader<D> {
[INFO] [stdout]    |        ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: fields `img_dir`, `transform`, and `target_transform` are never read
[INFO] [stdout]   --> src/nn/dataset.rs:33:5
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub struct CustomImageDataset {
[INFO] [stdout]    |            ------------------ fields in this struct
[INFO] [stdout] 32 |     img_labels: Vec<String>,
[INFO] [stdout] 33 |     img_dir: &'static str,
[INFO] [stdout]    |     ^^^^^^^
[INFO] [stdout] 34 |     transform: Option<Transform>,
[INFO] [stdout]    |     ^^^^^^^^^
[INFO] [stdout] 35 |     target_transform: Option<Transform>,
[INFO] [stdout]    |     ^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: multiple fields are never read
[INFO] [stdout]   --> src/nn/optimizer.rs:38:5
[INFO] [stdout]    |
[INFO] [stdout] 36 | pub struct Adam<T> {
[INFO] [stdout]    |            ---- fields in this struct
[INFO] [stdout] 37 |     /// Learning rate
[INFO] [stdout] 38 |     lr: T,
[INFO] [stdout]    |     ^^
[INFO] [stdout] 39 |     /// Momentum
[INFO] [stdout] 40 |     momentum: f32,
[INFO] [stdout]    |     ^^^^^^^^
[INFO] [stdout] 41 |     /// Decay rate
[INFO] [stdout] 42 |     decay_rate: f32,
[INFO] [stdout]    |     ^^^^^^^^^^
[INFO] [stdout] 43 |     /// Beta 1
[INFO] [stdout] 44 |     beta1: f32,
[INFO] [stdout]    |     ^^^^^
[INFO] [stdout] 45 |     /// Beta 2:
[INFO] [stdout] 46 |     beta2: f32,
[INFO] [stdout]    |     ^^^^^
[INFO] [stdout] 47 |     /// Epsilon
[INFO] [stdout] 48 |     epsilon: f32,
[INFO] [stdout]    |     ^^^^^^^
[INFO] [stdout] 49 |     m_dw: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 50 |     v_dw: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 51 |     m_db: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 52 |     v_db: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: fields `lr`, `momentum`, and `decay_rate` are never read
[INFO] [stdout]    --> src/nn/optimizer.rs:97:5
[INFO] [stdout]     |
[INFO] [stdout] 95  | pub struct SGD<T> {
[INFO] [stdout]     |            --- fields in this struct
[INFO] [stdout] 96  |     /// Learning rate
[INFO] [stdout] 97  |     lr: T,
[INFO] [stdout]     |     ^^
[INFO] [stdout] 98  |     /// Momentum
[INFO] [stdout] 99  |     momentum: f32,
[INFO] [stdout]     |     ^^^^^^^^
[INFO] [stdout] 100 |     /// Decay rate
[INFO] [stdout] 101 |     decay_rate: f32,
[INFO] [stdout]     |     ^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `swap` is never used
[INFO] [stdout]   --> src/tensor/mod.rs:33:4
[INFO] [stdout]    |
[INFO] [stdout] 33 | fn swap(lhs: &mut usize, rhs: &mut usize) {
[INFO] [stdout]    |    ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: methods `get_sub_tensor`, `sum`, and `prod` are never used
[INFO] [stdout]    --> src/tensor/mod.rs:686:8
[INFO] [stdout]     |
[INFO] [stdout] 560 | / impl<'a, T> Tensor<'a, T>
[INFO] [stdout] 561 | | where
[INFO] [stdout] 562 | |     T: TensorElement + Div<Output = T> + Sum<T>,
[INFO] [stdout] 563 | |     <T as FromStr>::Err: Error + 'static,
[INFO] [stdout] 564 | |     Vec<T>: IntoParallelIterator,
[INFO] [stdout] 565 | |     Vec<&'a T>: IntoParallelRefIterator<'a>,
[INFO] [stdout]     | |____________________________________________- methods in this implementation
[INFO] [stdout] ...
[INFO] [stdout] 686 |       fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec<T> {
[INFO] [stdout]     |          ^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout] 889 |       fn sum(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |          ^^^
[INFO] [stdout] ...
[INFO] [stdout] 904 |       fn prod(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |          ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `ReLU` should have a snake case name
[INFO] [stdout]   --> src/nn/activation.rs:25:8
[INFO] [stdout]    |
[INFO] [stdout] 25 | pub fn ReLU<'a, 'b, T>(x: &Tensor<'a, T>) -> Tensor<'b, T>
[INFO] [stdout]    |        ^^^^ help: convert the identifier to snake case: `re_lu`
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(non_snake_case)]` on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `PReLU` should have a snake case name
[INFO] [stdout]   --> src/nn/activation.rs:52:8
[INFO] [stdout]    |
[INFO] [stdout] 52 | pub fn PReLU<'a, 'b, T>(x: &Tensor<'a, T>, alpha: T) -> Tensor<'b, T>
[INFO] [stdout]    |        ^^^^^ help: convert the identifier to snake case: `pre_lu`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `Sigmoid` should have a snake case name
[INFO] [stdout]   --> src/nn/activation.rs:79:8
[INFO] [stdout]    |
[INFO] [stdout] 79 | pub fn Sigmoid<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case (notice the capitalization): `sigmoid`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `GeLU` should have a snake case name
[INFO] [stdout]    --> src/nn/activation.rs:108:8
[INFO] [stdout]     |
[INFO] [stdout] 108 | pub fn GeLU<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T>
[INFO] [stdout]     |        ^^^^ help: convert the identifier to snake case: `ge_lu`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `BCEntroypyLoss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:31:8
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^^^^^^^^^ help: convert the identifier to snake case: `bcentroypy_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `CEntroypyLoss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:50:8
[INFO] [stdout]    |
[INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^^^^^^^^ help: convert the identifier to snake case: `centroypy_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `L1Loss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:73:8
[INFO] [stdout]    |
[INFO] [stdout] 73 | pub fn L1Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^ help: convert the identifier to snake case: `l1_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `L2Loss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:96:8
[INFO] [stdout]    |
[INFO] [stdout] 96 | pub fn L2Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^ help: convert the identifier to snake case: `l2_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `MaxPool` should have a snake case name
[INFO] [stdout]   --> src/nn/pooling.rs:23:8
[INFO] [stdout]    |
[INFO] [stdout] 23 | pub fn MaxPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case: `max_pool`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `MinPool` should have a snake case name
[INFO] [stdout]   --> src/nn/pooling.rs:62:8
[INFO] [stdout]    |
[INFO] [stdout] 62 | pub fn MinPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case: `min_pool`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `AvgPool` should have a snake case name
[INFO] [stdout]   --> src/nn/pooling.rs:93:8
[INFO] [stdout]    |
[INFO] [stdout] 93 | pub fn AvgPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case: `avg_pool`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: 41 warnings emitted
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] running `Command { std: "docker" "inspect" "1cb98c66d491af650a3e8a835809619c3840a0a23092f05fa8af21c5e1f37292", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "1cb98c66d491af650a3e8a835809619c3840a0a23092f05fa8af21c5e1f37292", kill_on_drop: false }`
[INFO] [stdout] 1cb98c66d491af650a3e8a835809619c3840a0a23092f05fa8af21c5e1f37292
[INFO] checking kaffe-0.2.0 against try#9c7b5f5b6b8ec32bbbcf9f07fc0e5208da6c5f88 for pr-124157
[INFO] extracting crate kaffe 0.2.0 into /workspace/builds/worker-1-tc2/source
[INFO] validating manifest of crates.io crate kaffe 0.2.0 on toolchain 9c7b5f5b6b8ec32bbbcf9f07fc0e5208da6c5f88
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+9c7b5f5b6b8ec32bbbcf9f07fc0e5208da6c5f88" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }`
[INFO] started tweaking crates.io crate kaffe 0.2.0
[INFO] finished tweaking crates.io crate kaffe 0.2.0
[INFO] tweaked toml for crates.io crate kaffe 0.2.0 written to /workspace/builds/worker-1-tc2/source/Cargo.toml
[INFO] crate crates.io crate kaffe 0.2.0 already has a lockfile, it will not be regenerated
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+9c7b5f5b6b8ec32bbbcf9f07fc0e5208da6c5f88" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-1-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-1-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:99613afd962a8cfa530ec1899472a458bd015a1ab0af876cf7eb06f6006d81ea" "/opt/rustwide/cargo-home/bin/cargo" "+9c7b5f5b6b8ec32bbbcf9f07fc0e5208da6c5f88" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }`
[INFO] [stdout] fb261d71f6a1c261d65294636496b94922cf7acd8d698bb755584a889f30e420
[INFO] running `Command { std: "docker" "start" "-a" "fb261d71f6a1c261d65294636496b94922cf7acd8d698bb755584a889f30e420", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "inspect" "fb261d71f6a1c261d65294636496b94922cf7acd8d698bb755584a889f30e420", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "fb261d71f6a1c261d65294636496b94922cf7acd8d698bb755584a889f30e420", kill_on_drop: false }`
[INFO] [stdout] fb261d71f6a1c261d65294636496b94922cf7acd8d698bb755584a889f30e420
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-1-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-1-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:99613afd962a8cfa530ec1899472a458bd015a1ab0af876cf7eb06f6006d81ea" "/opt/rustwide/cargo-home/bin/cargo" "+9c7b5f5b6b8ec32bbbcf9f07fc0e5208da6c5f88" "check" "--frozen" "--all" "--all-targets" "--message-format=json", kill_on_drop: false }`
[INFO] [stdout] 96986c0fddeb32498a150113227c49ad5bb51fec6e9b38484d8a17605bb3624e
[INFO] running `Command { std: "docker" "start" "-a" "96986c0fddeb32498a150113227c49ad5bb51fec6e9b38484d8a17605bb3624e", kill_on_drop: false }`
[INFO] [stderr]    Compiling libc v0.2.146
[INFO] [stderr]    Compiling proc-macro2 v1.0.60
[INFO] [stderr]    Compiling quote v1.0.28
[INFO] [stderr]    Compiling unicode-ident v1.0.9
[INFO] [stderr]    Compiling crossbeam-utils v0.8.15
[INFO] [stderr]    Compiling serde v1.0.164
[INFO] [stderr]    Compiling num-traits v0.2.15
[INFO] [stderr]     Checking either v1.8.1
[INFO] [stderr]    Compiling rayon-core v1.11.0
[INFO] [stderr]    Compiling serde_json v1.0.96
[INFO] [stderr]    Compiling rustix v0.37.19
[INFO] [stderr]     Checking itertools v0.10.5
[INFO] [stderr]     Checking ciborium-io v0.2.1
[INFO] [stderr]     Checking anstyle v1.0.0
[INFO] [stderr]     Checking plotters-backend v0.3.4
[INFO] [stderr]     Checking linux-raw-sys v0.3.8
[INFO] [stderr]     Checking crossbeam-epoch v0.9.14
[INFO] [stderr]     Checking crossbeam-channel v0.5.8
[INFO] [stderr]    Compiling anyhow v1.0.71
[INFO] [stderr]     Checking crossbeam-deque v0.8.3
[INFO] [stderr]     Checking clap_lex v0.5.0
[INFO] [stderr]     Checking half v1.8.2
[INFO] [stderr]    Compiling syn v2.0.18
[INFO] [stderr]     Checking ryu v1.0.13
[INFO] [stderr]     Checking itoa v1.0.6
[INFO] [stderr]     Checking ciborium-ll v0.2.1
[INFO] [stderr]     Checking clap_builder v4.3.3
[INFO] [stderr]     Checking plotters-svg v0.3.3
[INFO] [stderr]     Checking regex-syntax v0.7.2
[INFO] [stderr]     Checking walkdir v2.3.3
[INFO] [stderr]     Checking once_cell v1.18.0
[INFO] [stderr]     Checking getrandom v0.2.10
[INFO] [stderr]     Checking num_cpus v1.15.0
[INFO] [stderr]     Checking io-lifetimes v1.0.11
[INFO] [stderr]     Checking plotters v0.3.4
[INFO] [stderr]     Checking rand_core v0.6.4
[INFO] [stderr]     Checking criterion-plot v0.5.0
[INFO] [stderr]     Checking rand_chacha v0.3.1
[INFO] [stderr]     Checking rand v0.8.5
[INFO] [stderr]     Checking rayon v1.7.0
[INFO] [stderr]     Checking regex v1.8.4
[INFO] [stderr]     Checking is-terminal v0.4.7
[INFO] [stderr]     Checking clap v4.3.3
[INFO] [stderr]    Compiling serde_derive v1.0.164
[INFO] [stderr]     Checking ciborium v0.2.1
[INFO] [stderr]     Checking kaffe v0.2.0 (/opt/rustwide/workdir)
[INFO] [stdout] warning: unused import: `marker::PhantomData`
[INFO] [stdout]  --> src/nn/optimizer.rs:4:25
[INFO] [stdout]   |
[INFO] [stdout] 4 | use std::{error::Error, marker::PhantomData, str::FromStr};
[INFO] [stdout]   |                         ^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout]   = note: `#[warn(unused_imports)]` on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `rayon::prelude::*`
[INFO] [stdout]  --> src/nn/optimizer.rs:7:5
[INFO] [stdout]   |
[INFO] [stdout] 7 | use rayon::prelude::*;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `idx`
[INFO] [stdout]   --> src/nn/dataset.rs:57:26
[INFO] [stdout]    |
[INFO] [stdout] 57 |     fn get<'a, T>(&self, idx: usize) -> (Tensor<'a, T>, String)
[INFO] [stdout]    |                          ^^^ help: if this is intentional, prefix it with an underscore: `_idx`
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(unused_variables)]` on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y`
[INFO] [stdout]   --> src/nn/loss.rs:31:30
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                              ^ help: if this is intentional, prefix it with an underscore: `_y`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y_hat`
[INFO] [stdout]   --> src/nn/loss.rs:31:49
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                                                 ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y`
[INFO] [stdout]   --> src/nn/loss.rs:50:29
[INFO] [stdout]    |
[INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                             ^ help: if this is intentional, prefix it with an underscore: `_y`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y_hat`
[INFO] [stdout]   --> src/nn/loss.rs:50:48
[INFO] [stdout]    |
[INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                                                ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `cost`
[INFO] [stdout]   --> src/nn/optimizer.rs:75:31
[INFO] [stdout]    |
[INFO] [stdout] 75 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]    |                               ^^^^ help: if this is intentional, prefix it with an underscore: `_cost`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `vars`
[INFO] [stdout]   --> src/nn/optimizer.rs:75:40
[INFO] [stdout]    |
[INFO] [stdout] 75 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]    |                                        ^^^^ help: if this is intentional, prefix it with an underscore: `_vars`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `cost`
[INFO] [stdout]    --> src/nn/optimizer.rs:117:31
[INFO] [stdout]     |
[INFO] [stdout] 117 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]     |                               ^^^^ help: if this is intentional, prefix it with an underscore: `_cost`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `vars`
[INFO] [stdout]    --> src/nn/optimizer.rs:117:40
[INFO] [stdout]     |
[INFO] [stdout] 117 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]     |                                        ^^^^ help: if this is intentional, prefix it with an underscore: `_vars`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `tensor`
[INFO] [stdout]   --> src/nn/mod.rs:24:5
[INFO] [stdout]    |
[INFO] [stdout] 24 |     tensor: Tensor<'a, T>,
[INFO] [stdout]    |     ^^^^^^ help: if this is intentional, prefix it with an underscore: `_tensor`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `kernel`
[INFO] [stdout]   --> src/nn/mod.rs:25:5
[INFO] [stdout]    |
[INFO] [stdout] 25 |     kernel: Tensor<'a, T>,
[INFO] [stdout]    |     ^^^^^^ help: if this is intentional, prefix it with an underscore: `_kernel`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `stride`
[INFO] [stdout]   --> src/nn/mod.rs:26:5
[INFO] [stdout]    |
[INFO] [stdout] 26 |     stride: usize,
[INFO] [stdout]    |     ^^^^^^ help: if this is intentional, prefix it with an underscore: `_stride`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `padding`
[INFO] [stdout]   --> src/nn/mod.rs:27:5
[INFO] [stdout]    |
[INFO] [stdout] 27 |     padding: usize,
[INFO] [stdout]    |     ^^^^^^^ help: if this is intentional, prefix it with an underscore: `_padding`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `f`
[INFO] [stdout]    --> src/tensor/mod.rs:257:19
[INFO] [stdout]     |
[INFO] [stdout] 257 |     fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
[INFO] [stdout]     |                   ^ help: if this is intentional, prefix it with an underscore: `_f`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `start_idx`
[INFO] [stdout]    --> src/tensor/mod.rs:686:30
[INFO] [stdout]     |
[INFO] [stdout] 686 |     fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec<T> {
[INFO] [stdout]     |                              ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_start_idx`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `size`
[INFO] [stdout]    --> src/tensor/mod.rs:686:48
[INFO] [stdout]     |
[INFO] [stdout] 686 |     fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec<T> {
[INFO] [stdout]     |                                                ^^^^ help: if this is intentional, prefix it with an underscore: `_size`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `rowcol`
[INFO] [stdout]    --> src/tensor/mod.rs:889:19
[INFO] [stdout]     |
[INFO] [stdout] 889 |     fn sum(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                   ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `dimension`
[INFO] [stdout]    --> src/tensor/mod.rs:889:34
[INFO] [stdout]     |
[INFO] [stdout] 889 |     fn sum(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                                  ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `rowcol`
[INFO] [stdout]    --> src/tensor/mod.rs:904:20
[INFO] [stdout]     |
[INFO] [stdout] 904 |     fn prod(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                    ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `dimension`
[INFO] [stdout]    --> src/tensor/mod.rs:904:35
[INFO] [stdout]     |
[INFO] [stdout] 904 |     fn prod(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                                   ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: field `layers` is never read
[INFO] [stdout]   --> src/nn/mod.rs:39:5
[INFO] [stdout]    |
[INFO] [stdout] 37 | pub struct Net {
[INFO] [stdout]    |            --- field in this struct
[INFO] [stdout] 38 |     /// Represents all layers in the network
[INFO] [stdout] 39 |     layers: Vec<Layer>,
[INFO] [stdout]    |     ^^^^^^
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(dead_code)]` on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: fields `data`, `batch_size`, and `shuffle` are never read
[INFO] [stdout]   --> src/nn/dataloader.rs:10:5
[INFO] [stdout]    |
[INFO] [stdout] 8  | pub struct DataLoader<D: DataSet + Clone> {
[INFO] [stdout]    |            ---------- fields in this struct
[INFO] [stdout] 9  |     /// Dataset
[INFO] [stdout] 10 |     data: D,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 11 |     /// Batch size
[INFO] [stdout] 12 |     batch_size: usize,
[INFO] [stdout]    |     ^^^^^^^^^^
[INFO] [stdout] 13 |     /// Whether or not to shuffle data
[INFO] [stdout] 14 |     shuffle: bool,
[INFO] [stdout]    |     ^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout]    = note: `DataLoader` has a derived impl for the trait `Clone`, but this is intentionally ignored during dead code analysis
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated items `new` and `iter` are never used
[INFO] [stdout]   --> src/nn/dataloader.rs:41:8
[INFO] [stdout]    |
[INFO] [stdout] 33 | impl<D: DataSet + Clone> DataLoader<D> {
[INFO] [stdout]    | -------------------------------------- associated items in this implementation
[INFO] [stdout] ...
[INFO] [stdout] 41 |     fn new(data: D, batch_size: usize, shuffle: bool) -> DataLoader<D> {
[INFO] [stdout]    |        ^^^
[INFO] [stdout] ...
[INFO] [stdout] 56 |     fn iter(&mut self) -> DataLoader<D> {
[INFO] [stdout]    |        ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: fields `img_dir`, `transform`, and `target_transform` are never read
[INFO] [stdout]   --> src/nn/dataset.rs:33:5
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub struct CustomImageDataset {
[INFO] [stdout]    |            ------------------ fields in this struct
[INFO] [stdout] 32 |     img_labels: Vec<String>,
[INFO] [stdout] 33 |     img_dir: &'static str,
[INFO] [stdout]    |     ^^^^^^^
[INFO] [stdout] 34 |     transform: Option<Transform>,
[INFO] [stdout]    |     ^^^^^^^^^
[INFO] [stdout] 35 |     target_transform: Option<Transform>,
[INFO] [stdout]    |     ^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: multiple fields are never read
[INFO] [stdout]   --> src/nn/optimizer.rs:38:5
[INFO] [stdout]    |
[INFO] [stdout] 36 | pub struct Adam<T> {
[INFO] [stdout]    |            ---- fields in this struct
[INFO] [stdout] 37 |     /// Learning rate
[INFO] [stdout] 38 |     lr: T,
[INFO] [stdout]    |     ^^
[INFO] [stdout] 39 |     /// Momentum
[INFO] [stdout] 40 |     momentum: f32,
[INFO] [stdout]    |     ^^^^^^^^
[INFO] [stdout] 41 |     /// Decay rate
[INFO] [stdout] 42 |     decay_rate: f32,
[INFO] [stdout]    |     ^^^^^^^^^^
[INFO] [stdout] 43 |     /// Beta 1
[INFO] [stdout] 44 |     beta1: f32,
[INFO] [stdout]    |     ^^^^^
[INFO] [stdout] 45 |     /// Beta 2:
[INFO] [stdout] 46 |     beta2: f32,
[INFO] [stdout]    |     ^^^^^
[INFO] [stdout] 47 |     /// Epsilon
[INFO] [stdout] 48 |     epsilon: f32,
[INFO] [stdout]    |     ^^^^^^^
[INFO] [stdout] 49 |     m_dw: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 50 |     v_dw: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 51 |     m_db: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 52 |     v_db: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: fields `lr`, `momentum`, and `decay_rate` are never read
[INFO] [stdout]    --> src/nn/optimizer.rs:97:5
[INFO] [stdout]     |
[INFO] [stdout] 95  | pub struct SGD<T> {
[INFO] [stdout]     |            --- fields in this struct
[INFO] [stdout] 96  |     /// Learning rate
[INFO] [stdout] 97  |     lr: T,
[INFO] [stdout]     |     ^^
[INFO] [stdout] 98  |     /// Momentum
[INFO] [stdout] 99  |     momentum: f32,
[INFO] [stdout]     |     ^^^^^^^^
[INFO] [stdout] 100 |     /// Decay rate
[INFO] [stdout] 101 |     decay_rate: f32,
[INFO] [stdout]     |     ^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `swap` is never used
[INFO] [stdout]   --> src/tensor/mod.rs:33:4
[INFO] [stdout]    |
[INFO] [stdout] 33 | fn swap(lhs: &mut usize, rhs: &mut usize) {
[INFO] [stdout]    |    ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: methods `get_sub_tensor`, `sum`, and `prod` are never used
[INFO] [stdout]    --> src/tensor/mod.rs:686:8
[INFO] [stdout]     |
[INFO] [stdout] 560 | / impl<'a, T> Tensor<'a, T>
[INFO] [stdout] 561 | | where
[INFO] [stdout] 562 | |     T: TensorElement + Div<Output = T> + Sum<T>,
[INFO] [stdout] 563 | |     <T as FromStr>::Err: Error + 'static,
[INFO] [stdout] 564 | |     Vec<T>: IntoParallelIterator,
[INFO] [stdout] 565 | |     Vec<&'a T>: IntoParallelRefIterator<'a>,
[INFO] [stdout]     | |____________________________________________- methods in this implementation
[INFO] [stdout] ...
[INFO] [stdout] 686 |       fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec<T> {
[INFO] [stdout]     |          ^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout] 889 |       fn sum(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |          ^^^
[INFO] [stdout] ...
[INFO] [stdout] 904 |       fn prod(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |          ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `ReLU` should have a snake case name
[INFO] [stdout]   --> src/nn/activation.rs:25:8
[INFO] [stdout]    |
[INFO] [stdout] 25 | pub fn ReLU<'a, 'b, T>(x: &Tensor<'a, T>) -> Tensor<'b, T>
[INFO] [stdout]    |        ^^^^ help: convert the identifier to snake case: `re_lu`
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(non_snake_case)]` on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `PReLU` should have a snake case name
[INFO] [stdout]   --> src/nn/activation.rs:52:8
[INFO] [stdout]    |
[INFO] [stdout] 52 | pub fn PReLU<'a, 'b, T>(x: &Tensor<'a, T>, alpha: T) -> Tensor<'b, T>
[INFO] [stdout]    |        ^^^^^ help: convert the identifier to snake case: `pre_lu`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `Sigmoid` should have a snake case name
[INFO] [stdout]   --> src/nn/activation.rs:79:8
[INFO] [stdout]    |
[INFO] [stdout] 79 | pub fn Sigmoid<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case (notice the capitalization): `sigmoid`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `GeLU` should have a snake case name
[INFO] [stdout]    --> src/nn/activation.rs:108:8
[INFO] [stdout]     |
[INFO] [stdout] 108 | pub fn GeLU<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T>
[INFO] [stdout]     |        ^^^^ help: convert the identifier to snake case: `ge_lu`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: missing documentation for an associated function
[INFO] [stdout]   --> src/nn/loss.rs:14:5
[INFO] [stdout]    |
[INFO] [stdout] 14 |     fn loss(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T;
[INFO] [stdout]    |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout] note: the lint level is defined here
[INFO] [stdout]   --> src/nn/loss.rs:2:9
[INFO] [stdout]    |
[INFO] [stdout] 2  | #![warn(missing_docs)]
[INFO] [stdout]    |         ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: missing documentation for an associated function
[INFO] [stdout]   --> src/nn/loss.rs:15:5
[INFO] [stdout]    |
[INFO] [stdout] 15 |     fn backward();
[INFO] [stdout]    |     ^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `BCEntroypyLoss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:31:8
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^^^^^^^^^ help: convert the identifier to snake case: `bcentroypy_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `CEntroypyLoss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:50:8
[INFO] [stdout]    |
[INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^^^^^^^^ help: convert the identifier to snake case: `centroypy_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `L1Loss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:73:8
[INFO] [stdout]    |
[INFO] [stdout] 73 | pub fn L1Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^ help: convert the identifier to snake case: `l1_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `L2Loss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:96:8
[INFO] [stdout]    |
[INFO] [stdout] 96 | pub fn L2Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^ help: convert the identifier to snake case: `l2_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `MaxPool` should have a snake case name
[INFO] [stdout]   --> src/nn/pooling.rs:23:8
[INFO] [stdout]    |
[INFO] [stdout] 23 | pub fn MaxPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case: `max_pool`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `MinPool` should have a snake case name
[INFO] [stdout]   --> src/nn/pooling.rs:62:8
[INFO] [stdout]    |
[INFO] [stdout] 62 | pub fn MinPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case: `min_pool`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `AvgPool` should have a snake case name
[INFO] [stdout]   --> src/nn/pooling.rs:93:8
[INFO] [stdout]    |
[INFO] [stdout] 93 | pub fn AvgPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case: `avg_pool`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: missing documentation for a struct
[INFO] [stdout]  --> src/nn/transform.rs:4:1
[INFO] [stdout]   |
[INFO] [stdout] 4 | pub struct Transform;
[INFO] [stdout]   | ^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout] note: the lint level is defined here
[INFO] [stdout]  --> src/nn/transform.rs:2:9
[INFO] [stdout]   |
[INFO] [stdout] 2 | #![warn(missing_docs)]
[INFO] [stdout]   |         ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: missing documentation for a struct
[INFO] [stdout]    --> src/tensor/mod.rs:145:1
[INFO] [stdout]     |
[INFO] [stdout] 145 | pub struct Tensor<'a, T>
[INFO] [stdout]     | ^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] note: the lint level is defined here
[INFO] [stdout]    --> src/lib.rs:1:9
[INFO] [stdout]     |
[INFO] [stdout] 1   | #![warn(missing_docs)]
[INFO] [stdout]     |         ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: 45 warnings emitted
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stderr]     Checking tinytemplate v1.2.1
[INFO] [stderr]     Checking criterion v0.5.1
[INFO] [stdout] error[E0432]: unresolved imports `kaffe::Matrix`, `kaffe::MatrixLinAlg`
[INFO] [stdout]  --> benches/matmul_benchmark.rs:1:13
[INFO] [stdout]   |
[INFO] [stdout] 1 | use kaffe::{Matrix, MatrixLinAlg};
[INFO] [stdout]   |             ^^^^^^  ^^^^^^^^^^^^ no `MatrixLinAlg` in the root
[INFO] [stdout]   |             |
[INFO] [stdout]   |             no `Matrix` in the root
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: aborting due to 1 previous error
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] For more information about this error, try `rustc --explain E0432`.
[INFO] [stdout] 
[INFO] [stderr] error: could not compile `kaffe` (bench "matmul_benchmark") due to 2 previous errors
[INFO] [stderr] warning: build failed, waiting for other jobs to finish...
[INFO] [stdout] warning: unused import: `marker::PhantomData`
[INFO] [stdout]  --> src/nn/optimizer.rs:4:25
[INFO] [stdout]   |
[INFO] [stdout] 4 | use std::{error::Error, marker::PhantomData, str::FromStr};
[INFO] [stdout]   |                         ^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout]   = note: `#[warn(unused_imports)]` on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `rayon::prelude::*`
[INFO] [stdout]  --> src/nn/optimizer.rs:7:5
[INFO] [stdout]   |
[INFO] [stdout] 7 | use rayon::prelude::*;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0432]: unresolved imports `kaffe::Matrix`, `kaffe::MatrixLinAlg`
[INFO] [stdout]  --> benches/matrix_ops_benches.rs:1:13
[INFO] [stdout]   |
[INFO] [stdout] 1 | use kaffe::{Matrix, MatrixLinAlg};
[INFO] [stdout]   |             ^^^^^^  ^^^^^^^^^^^^ no `MatrixLinAlg` in the root
[INFO] [stdout]   |             |
[INFO] [stdout]   |             no `Matrix` in the root
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: aborting due to 1 previous error
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] For more information about this error, try `rustc --explain E0432`.
[INFO] [stdout] 
[INFO] [stderr] error: could not compile `kaffe` (bench "matrix_ops_benches") due to 2 previous errors
[INFO] [stdout] warning: unused variable: `idx`
[INFO] [stdout]   --> src/nn/dataset.rs:57:26
[INFO] [stdout]    |
[INFO] [stdout] 57 |     fn get<'a, T>(&self, idx: usize) -> (Tensor<'a, T>, String)
[INFO] [stdout]    |                          ^^^ help: if this is intentional, prefix it with an underscore: `_idx`
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(unused_variables)]` on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y`
[INFO] [stdout]   --> src/nn/loss.rs:31:30
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                              ^ help: if this is intentional, prefix it with an underscore: `_y`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y_hat`
[INFO] [stdout]   --> src/nn/loss.rs:31:49
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                                                 ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y`
[INFO] [stdout]   --> src/nn/loss.rs:50:29
[INFO] [stdout]    |
[INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                             ^ help: if this is intentional, prefix it with an underscore: `_y`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `y_hat`
[INFO] [stdout]   --> src/nn/loss.rs:50:48
[INFO] [stdout]    |
[INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |                                                ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `cost`
[INFO] [stdout]   --> src/nn/optimizer.rs:75:31
[INFO] [stdout]    |
[INFO] [stdout] 75 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]    |                               ^^^^ help: if this is intentional, prefix it with an underscore: `_cost`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `vars`
[INFO] [stdout]   --> src/nn/optimizer.rs:75:40
[INFO] [stdout]    |
[INFO] [stdout] 75 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]    |                                        ^^^^ help: if this is intentional, prefix it with an underscore: `_vars`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `cost`
[INFO] [stdout]    --> src/nn/optimizer.rs:117:31
[INFO] [stdout]     |
[INFO] [stdout] 117 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]     |                               ^^^^ help: if this is intentional, prefix it with an underscore: `_cost`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `vars`
[INFO] [stdout]    --> src/nn/optimizer.rs:117:40
[INFO] [stdout]     |
[INFO] [stdout] 117 |     fn minimize<F>(&mut self, cost: F, vars: &mut Vec<T>)
[INFO] [stdout]     |                                        ^^^^ help: if this is intentional, prefix it with an underscore: `_vars`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `tensor`
[INFO] [stdout]   --> src/nn/mod.rs:24:5
[INFO] [stdout]    |
[INFO] [stdout] 24 |     tensor: Tensor<'a, T>,
[INFO] [stdout]    |     ^^^^^^ help: if this is intentional, prefix it with an underscore: `_tensor`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `kernel`
[INFO] [stdout]   --> src/nn/mod.rs:25:5
[INFO] [stdout]    |
[INFO] [stdout] 25 |     kernel: Tensor<'a, T>,
[INFO] [stdout]    |     ^^^^^^ help: if this is intentional, prefix it with an underscore: `_kernel`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `stride`
[INFO] [stdout]   --> src/nn/mod.rs:26:5
[INFO] [stdout]    |
[INFO] [stdout] 26 |     stride: usize,
[INFO] [stdout]    |     ^^^^^^ help: if this is intentional, prefix it with an underscore: `_stride`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `padding`
[INFO] [stdout]   --> src/nn/mod.rs:27:5
[INFO] [stdout]    |
[INFO] [stdout] 27 |     padding: usize,
[INFO] [stdout]    |     ^^^^^^^ help: if this is intentional, prefix it with an underscore: `_padding`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `f`
[INFO] [stdout]    --> src/tensor/mod.rs:257:19
[INFO] [stdout]     |
[INFO] [stdout] 257 |     fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
[INFO] [stdout]     |                   ^ help: if this is intentional, prefix it with an underscore: `_f`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `start_idx`
[INFO] [stdout]    --> src/tensor/mod.rs:686:30
[INFO] [stdout]     |
[INFO] [stdout] 686 |     fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec<T> {
[INFO] [stdout]     |                              ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_start_idx`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `size`
[INFO] [stdout]    --> src/tensor/mod.rs:686:48
[INFO] [stdout]     |
[INFO] [stdout] 686 |     fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec<T> {
[INFO] [stdout]     |                                                ^^^^ help: if this is intentional, prefix it with an underscore: `_size`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `rowcol`
[INFO] [stdout]    --> src/tensor/mod.rs:889:19
[INFO] [stdout]     |
[INFO] [stdout] 889 |     fn sum(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                   ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `dimension`
[INFO] [stdout]    --> src/tensor/mod.rs:889:34
[INFO] [stdout]     |
[INFO] [stdout] 889 |     fn sum(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                                  ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `rowcol`
[INFO] [stdout]    --> src/tensor/mod.rs:904:20
[INFO] [stdout]     |
[INFO] [stdout] 904 |     fn prod(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                    ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `dimension`
[INFO] [stdout]    --> src/tensor/mod.rs:904:35
[INFO] [stdout]     |
[INFO] [stdout] 904 |     fn prod(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |                                   ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: field `layers` is never read
[INFO] [stdout]   --> src/nn/mod.rs:39:5
[INFO] [stdout]    |
[INFO] [stdout] 37 | pub struct Net {
[INFO] [stdout]    |            --- field in this struct
[INFO] [stdout] 38 |     /// Represents all layers in the network
[INFO] [stdout] 39 |     layers: Vec<Layer>,
[INFO] [stdout]    |     ^^^^^^
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(dead_code)]` on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: fields `data`, `batch_size`, and `shuffle` are never read
[INFO] [stdout]   --> src/nn/dataloader.rs:10:5
[INFO] [stdout]    |
[INFO] [stdout] 8  | pub struct DataLoader<D: DataSet + Clone> {
[INFO] [stdout]    |            ---------- fields in this struct
[INFO] [stdout] 9  |     /// Dataset
[INFO] [stdout] 10 |     data: D,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 11 |     /// Batch size
[INFO] [stdout] 12 |     batch_size: usize,
[INFO] [stdout]    |     ^^^^^^^^^^
[INFO] [stdout] 13 |     /// Whether or not to shuffle data
[INFO] [stdout] 14 |     shuffle: bool,
[INFO] [stdout]    |     ^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout]    = note: `DataLoader` has a derived impl for the trait `Clone`, but this is intentionally ignored during dead code analysis
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated items `new` and `iter` are never used
[INFO] [stdout]   --> src/nn/dataloader.rs:41:8
[INFO] [stdout]    |
[INFO] [stdout] 33 | impl<D: DataSet + Clone> DataLoader<D> {
[INFO] [stdout]    | -------------------------------------- associated items in this implementation
[INFO] [stdout] ...
[INFO] [stdout] 41 |     fn new(data: D, batch_size: usize, shuffle: bool) -> DataLoader<D> {
[INFO] [stdout]    |        ^^^
[INFO] [stdout] ...
[INFO] [stdout] 56 |     fn iter(&mut self) -> DataLoader<D> {
[INFO] [stdout]    |        ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: fields `img_dir`, `transform`, and `target_transform` are never read
[INFO] [stdout]   --> src/nn/dataset.rs:33:5
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub struct CustomImageDataset {
[INFO] [stdout]    |            ------------------ fields in this struct
[INFO] [stdout] 32 |     img_labels: Vec<String>,
[INFO] [stdout] 33 |     img_dir: &'static str,
[INFO] [stdout]    |     ^^^^^^^
[INFO] [stdout] 34 |     transform: Option<Transform>,
[INFO] [stdout]    |     ^^^^^^^^^
[INFO] [stdout] 35 |     target_transform: Option<Transform>,
[INFO] [stdout]    |     ^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: multiple fields are never read
[INFO] [stdout]   --> src/nn/optimizer.rs:38:5
[INFO] [stdout]    |
[INFO] [stdout] 36 | pub struct Adam<T> {
[INFO] [stdout]    |            ---- fields in this struct
[INFO] [stdout] 37 |     /// Learning rate
[INFO] [stdout] 38 |     lr: T,
[INFO] [stdout]    |     ^^
[INFO] [stdout] 39 |     /// Momentum
[INFO] [stdout] 40 |     momentum: f32,
[INFO] [stdout]    |     ^^^^^^^^
[INFO] [stdout] 41 |     /// Decay rate
[INFO] [stdout] 42 |     decay_rate: f32,
[INFO] [stdout]    |     ^^^^^^^^^^
[INFO] [stdout] 43 |     /// Beta 1
[INFO] [stdout] 44 |     beta1: f32,
[INFO] [stdout]    |     ^^^^^
[INFO] [stdout] 45 |     /// Beta 2:
[INFO] [stdout] 46 |     beta2: f32,
[INFO] [stdout]    |     ^^^^^
[INFO] [stdout] 47 |     /// Epsilon
[INFO] [stdout] 48 |     epsilon: f32,
[INFO] [stdout]    |     ^^^^^^^
[INFO] [stdout] 49 |     m_dw: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 50 |     v_dw: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 51 |     m_db: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 52 |     v_db: f32,
[INFO] [stdout]    |     ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: fields `lr`, `momentum`, and `decay_rate` are never read
[INFO] [stdout]    --> src/nn/optimizer.rs:97:5
[INFO] [stdout]     |
[INFO] [stdout] 95  | pub struct SGD<T> {
[INFO] [stdout]     |            --- fields in this struct
[INFO] [stdout] 96  |     /// Learning rate
[INFO] [stdout] 97  |     lr: T,
[INFO] [stdout]     |     ^^
[INFO] [stdout] 98  |     /// Momentum
[INFO] [stdout] 99  |     momentum: f32,
[INFO] [stdout]     |     ^^^^^^^^
[INFO] [stdout] 100 |     /// Decay rate
[INFO] [stdout] 101 |     decay_rate: f32,
[INFO] [stdout]     |     ^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `swap` is never used
[INFO] [stdout]   --> src/tensor/mod.rs:33:4
[INFO] [stdout]    |
[INFO] [stdout] 33 | fn swap(lhs: &mut usize, rhs: &mut usize) {
[INFO] [stdout]    |    ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: methods `get_sub_tensor`, `sum`, and `prod` are never used
[INFO] [stdout]    --> src/tensor/mod.rs:686:8
[INFO] [stdout]     |
[INFO] [stdout] 560 | / impl<'a, T> Tensor<'a, T>
[INFO] [stdout] 561 | | where
[INFO] [stdout] 562 | |     T: TensorElement + Div<Output = T> + Sum<T>,
[INFO] [stdout] 563 | |     <T as FromStr>::Err: Error + 'static,
[INFO] [stdout] 564 | |     Vec<T>: IntoParallelIterator,
[INFO] [stdout] 565 | |     Vec<&'a T>: IntoParallelRefIterator<'a>,
[INFO] [stdout]     | |____________________________________________- methods in this implementation
[INFO] [stdout] ...
[INFO] [stdout] 686 |       fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec<T> {
[INFO] [stdout]     |          ^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout] 889 |       fn sum(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |          ^^^
[INFO] [stdout] ...
[INFO] [stdout] 904 |       fn prod(&self, rowcol: usize, dimension: Dimension) -> T {
[INFO] [stdout]     |          ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `ReLU` should have a snake case name
[INFO] [stdout]   --> src/nn/activation.rs:25:8
[INFO] [stdout]    |
[INFO] [stdout] 25 | pub fn ReLU<'a, 'b, T>(x: &Tensor<'a, T>) -> Tensor<'b, T>
[INFO] [stdout]    |        ^^^^ help: convert the identifier to snake case: `re_lu`
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(non_snake_case)]` on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `PReLU` should have a snake case name
[INFO] [stdout]   --> src/nn/activation.rs:52:8
[INFO] [stdout]    |
[INFO] [stdout] 52 | pub fn PReLU<'a, 'b, T>(x: &Tensor<'a, T>, alpha: T) -> Tensor<'b, T>
[INFO] [stdout]    |        ^^^^^ help: convert the identifier to snake case: `pre_lu`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `Sigmoid` should have a snake case name
[INFO] [stdout]   --> src/nn/activation.rs:79:8
[INFO] [stdout]    |
[INFO] [stdout] 79 | pub fn Sigmoid<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case (notice the capitalization): `sigmoid`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `GeLU` should have a snake case name
[INFO] [stdout]    --> src/nn/activation.rs:108:8
[INFO] [stdout]     |
[INFO] [stdout] 108 | pub fn GeLU<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T>
[INFO] [stdout]     |        ^^^^ help: convert the identifier to snake case: `ge_lu`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `BCEntroypyLoss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:31:8
[INFO] [stdout]    |
[INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^^^^^^^^^ help: convert the identifier to snake case: `bcentroypy_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `CEntroypyLoss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:50:8
[INFO] [stdout]    |
[INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^^^^^^^^ help: convert the identifier to snake case: `centroypy_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `L1Loss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:73:8
[INFO] [stdout]    |
[INFO] [stdout] 73 | pub fn L1Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^ help: convert the identifier to snake case: `l1_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `L2Loss` should have a snake case name
[INFO] [stdout]   --> src/nn/loss.rs:96:8
[INFO] [stdout]    |
[INFO] [stdout] 96 | pub fn L2Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T
[INFO] [stdout]    |        ^^^^^^ help: convert the identifier to snake case: `l2_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `MaxPool` should have a snake case name
[INFO] [stdout]   --> src/nn/pooling.rs:23:8
[INFO] [stdout]    |
[INFO] [stdout] 23 | pub fn MaxPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case: `max_pool`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `MinPool` should have a snake case name
[INFO] [stdout]   --> src/nn/pooling.rs:62:8
[INFO] [stdout]    |
[INFO] [stdout] 62 | pub fn MinPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case: `min_pool`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `AvgPool` should have a snake case name
[INFO] [stdout]   --> src/nn/pooling.rs:93:8
[INFO] [stdout]    |
[INFO] [stdout] 93 | pub fn AvgPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T>
[INFO] [stdout]    |        ^^^^^^^ help: convert the identifier to snake case: `avg_pool`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: 41 warnings emitted
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] running `Command { std: "docker" "inspect" "96986c0fddeb32498a150113227c49ad5bb51fec6e9b38484d8a17605bb3624e", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "96986c0fddeb32498a150113227c49ad5bb51fec6e9b38484d8a17605bb3624e", kill_on_drop: false }`
[INFO] [stdout] 96986c0fddeb32498a150113227c49ad5bb51fec6e9b38484d8a17605bb3624e
