[INFO] fetching crate kaffe 0.2.0... [INFO] testing kaffe-0.2.0 against try#c2e32f1c9652b13ed99608599c1e855462f421f3 for pr-146098-7 [INFO] extracting crate kaffe 0.2.0 into /workspace/builds/worker-5-tc2/source [INFO] started tweaking crates.io crate kaffe 0.2.0 [INFO] finished tweaking crates.io crate kaffe 0.2.0 [INFO] tweaked toml for crates.io crate kaffe 0.2.0 written to /workspace/builds/worker-5-tc2/source/Cargo.toml [INFO] validating manifest of crates.io crate kaffe 0.2.0 on toolchain c2e32f1c9652b13ed99608599c1e855462f421f3 [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+c2e32f1c9652b13ed99608599c1e855462f421f3" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }` [INFO] crate crates.io crate kaffe 0.2.0 already has a lockfile, it will not be regenerated [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+c2e32f1c9652b13ed99608599c1e855462f421f3" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }` [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:4848fb76d95f26979359cc7e45710b1dbc8f3acb7aeedee7c460d7702230f228" "/opt/rustwide/cargo-home/bin/cargo" "+c2e32f1c9652b13ed99608599c1e855462f421f3" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }` [INFO] [stdout] 19f7e7a0787420a79767098147a66b774b63f4b831ee58f5cc831e7f63e75a4f [INFO] running `Command { std: "docker" "start" "-a" "19f7e7a0787420a79767098147a66b774b63f4b831ee58f5cc831e7f63e75a4f", kill_on_drop: false }` [INFO] running `Command { std: "docker" "inspect" "19f7e7a0787420a79767098147a66b774b63f4b831ee58f5cc831e7f63e75a4f", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "19f7e7a0787420a79767098147a66b774b63f4b831ee58f5cc831e7f63e75a4f", kill_on_drop: false }` [INFO] [stdout] 19f7e7a0787420a79767098147a66b774b63f4b831ee58f5cc831e7f63e75a4f [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:4848fb76d95f26979359cc7e45710b1dbc8f3acb7aeedee7c460d7702230f228" "/opt/rustwide/cargo-home/bin/cargo" "+c2e32f1c9652b13ed99608599c1e855462f421f3" "build" "--frozen" "--message-format=json", kill_on_drop: false }` [INFO] [stdout] 24cd81002956ca7d25a15907c53115dea16a1ba750654bc64b0367ed2ffbb797 [INFO] running `Command { std: "docker" "start" "-a" "24cd81002956ca7d25a15907c53115dea16a1ba750654bc64b0367ed2ffbb797", kill_on_drop: false }` [INFO] [stderr] Compiling autocfg v1.1.0 [INFO] [stderr] Compiling libc v0.2.146 [INFO] [stderr] Compiling crossbeam-utils v0.8.15 [INFO] [stderr] Compiling proc-macro2 v1.0.60 [INFO] [stderr] Compiling unicode-ident v1.0.9 [INFO] [stderr] Compiling quote v1.0.28 [INFO] [stderr] Compiling scopeguard v1.1.0 [INFO] [stderr] Compiling rayon-core v1.11.0 [INFO] [stderr] Compiling serde v1.0.164 [INFO] [stderr] Compiling ppv-lite86 v0.2.17 [INFO] [stderr] Compiling either v1.8.1 [INFO] [stderr] Compiling anyhow v1.0.71 [INFO] [stderr] Compiling itertools v0.10.5 [INFO] [stderr] Compiling memoffset v0.8.0 [INFO] [stderr] Compiling crossbeam-epoch v0.9.14 [INFO] [stderr] Compiling num-traits v0.2.15 [INFO] [stderr] Compiling syn v2.0.18 [INFO] [stderr] Compiling crossbeam-channel v0.5.8 [INFO] [stderr] Compiling crossbeam-deque v0.8.3 [INFO] [stderr] Compiling getrandom v0.2.10 [INFO] [stderr] Compiling num_cpus v1.15.0 [INFO] [stderr] Compiling rand_core v0.6.4 [INFO] [stderr] Compiling rand_chacha v0.3.1 [INFO] [stderr] Compiling rand v0.8.5 [INFO] [stderr] Compiling rayon v1.7.0 [INFO] [stderr] Compiling serde_derive v1.0.164 [INFO] [stderr] Compiling kaffe v0.2.0 (/opt/rustwide/workdir) [INFO] [stdout] warning: unused import: `marker::PhantomData` [INFO] [stdout] --> src/nn/optimizer.rs:4:25 [INFO] [stdout] | [INFO] [stdout] 4 | use std::{error::Error, marker::PhantomData, str::FromStr}; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `rayon::prelude::*` [INFO] [stdout] --> src/nn/optimizer.rs:7:5 [INFO] [stdout] | [INFO] [stdout] 7 | use rayon::prelude::*; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `idx` [INFO] [stdout] --> src/nn/dataset.rs:57:26 [INFO] [stdout] | [INFO] [stdout] 57 | fn get<'a, T>(&self, idx: usize) -> (Tensor<'a, T>, String) [INFO] [stdout] | ^^^ help: if this is intentional, prefix it with an underscore: `_idx` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_variables)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `y` [INFO] [stdout] --> src/nn/loss.rs:31:30 [INFO] [stdout] | [INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_y` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `y_hat` [INFO] [stdout] --> src/nn/loss.rs:31:49 [INFO] [stdout] | [INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `y` [INFO] [stdout] --> src/nn/loss.rs:50:29 [INFO] [stdout] | [INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_y` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `y_hat` [INFO] [stdout] --> src/nn/loss.rs:50:48 [INFO] [stdout] | [INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `cost` [INFO] [stdout] --> src/nn/optimizer.rs:75:31 [INFO] [stdout] | [INFO] [stdout] 75 | fn minimize(&mut self, cost: F, vars: &mut Vec) [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_cost` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `vars` [INFO] [stdout] --> src/nn/optimizer.rs:75:40 [INFO] [stdout] | [INFO] [stdout] 75 | fn minimize(&mut self, cost: F, vars: &mut Vec) [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_vars` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `cost` [INFO] [stdout] --> src/nn/optimizer.rs:117:31 [INFO] [stdout] | [INFO] [stdout] 117 | fn minimize(&mut self, cost: F, vars: &mut Vec) [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_cost` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `vars` [INFO] [stdout] --> src/nn/optimizer.rs:117:40 [INFO] [stdout] | [INFO] [stdout] 117 | fn minimize(&mut self, cost: F, vars: &mut Vec) [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_vars` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `tensor` [INFO] [stdout] --> src/nn/mod.rs:24:5 [INFO] [stdout] | [INFO] [stdout] 24 | tensor: Tensor<'a, T>, [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `kernel` [INFO] [stdout] --> src/nn/mod.rs:25:5 [INFO] [stdout] | [INFO] [stdout] 25 | kernel: Tensor<'a, T>, [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_kernel` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `stride` [INFO] [stdout] --> src/nn/mod.rs:26:5 [INFO] [stdout] | [INFO] [stdout] 26 | stride: usize, [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_stride` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `padding` [INFO] [stdout] --> src/nn/mod.rs:27:5 [INFO] [stdout] | [INFO] [stdout] 27 | padding: usize, [INFO] [stdout] | ^^^^^^^ help: if this is intentional, prefix it with an underscore: `_padding` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `f` [INFO] [stdout] --> src/tensor/mod.rs:257:19 [INFO] [stdout] | [INFO] [stdout] 257 | fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_f` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `start_idx` [INFO] [stdout] --> src/tensor/mod.rs:686:30 [INFO] [stdout] | [INFO] [stdout] 686 | fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec { [INFO] [stdout] | ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_start_idx` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `size` [INFO] [stdout] --> src/tensor/mod.rs:686:48 [INFO] [stdout] | [INFO] [stdout] 686 | fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec { [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_size` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `rowcol` [INFO] [stdout] --> src/tensor/mod.rs:889:19 [INFO] [stdout] | [INFO] [stdout] 889 | fn sum(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `dimension` [INFO] [stdout] --> src/tensor/mod.rs:889:34 [INFO] [stdout] | [INFO] [stdout] 889 | fn sum(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `rowcol` [INFO] [stdout] --> src/tensor/mod.rs:904:20 [INFO] [stdout] | [INFO] [stdout] 904 | fn prod(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `dimension` [INFO] [stdout] --> src/tensor/mod.rs:904:35 [INFO] [stdout] | [INFO] [stdout] 904 | fn prod(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `layers` is never read [INFO] [stdout] --> src/nn/mod.rs:39:5 [INFO] [stdout] | [INFO] [stdout] 37 | pub struct Net { [INFO] [stdout] | --- field in this struct [INFO] [stdout] 38 | /// Represents all layers in the network [INFO] [stdout] 39 | layers: Vec, [INFO] [stdout] | ^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: fields `data`, `batch_size`, and `shuffle` are never read [INFO] [stdout] --> src/nn/dataloader.rs:10:5 [INFO] [stdout] | [INFO] [stdout] 8 | pub struct DataLoader { [INFO] [stdout] | ---------- fields in this struct [INFO] [stdout] 9 | /// Dataset [INFO] [stdout] 10 | data: D, [INFO] [stdout] | ^^^^ [INFO] [stdout] 11 | /// Batch size [INFO] [stdout] 12 | batch_size: usize, [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] 13 | /// Whether or not to shuffle data [INFO] [stdout] 14 | shuffle: bool, [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `DataLoader` has a derived impl for the trait `Clone`, but this is intentionally ignored during dead code analysis [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: associated items `new` and `iter` are never used [INFO] [stdout] --> src/nn/dataloader.rs:41:8 [INFO] [stdout] | [INFO] [stdout] 33 | impl DataLoader { [INFO] [stdout] | -------------------------------------- associated items in this implementation [INFO] [stdout] ... [INFO] [stdout] 41 | fn new(data: D, batch_size: usize, shuffle: bool) -> DataLoader { [INFO] [stdout] | ^^^ [INFO] [stdout] ... [INFO] [stdout] 56 | fn iter(&mut self) -> DataLoader { [INFO] [stdout] | ^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: fields `img_dir`, `transform`, and `target_transform` are never read [INFO] [stdout] --> src/nn/dataset.rs:33:5 [INFO] [stdout] | [INFO] [stdout] 31 | pub struct CustomImageDataset { [INFO] [stdout] | ------------------ fields in this struct [INFO] [stdout] 32 | img_labels: Vec, [INFO] [stdout] 33 | img_dir: &'static str, [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] 34 | transform: Option, [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] 35 | target_transform: Option, [INFO] [stdout] | ^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: multiple fields are never read [INFO] [stdout] --> src/nn/optimizer.rs:38:5 [INFO] [stdout] | [INFO] [stdout] 36 | pub struct Adam { [INFO] [stdout] | ---- fields in this struct [INFO] [stdout] 37 | /// Learning rate [INFO] [stdout] 38 | lr: T, [INFO] [stdout] | ^^ [INFO] [stdout] 39 | /// Momentum [INFO] [stdout] 40 | momentum: f32, [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] 41 | /// Decay rate [INFO] [stdout] 42 | decay_rate: f32, [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] 43 | /// Beta 1 [INFO] [stdout] 44 | beta1: f32, [INFO] [stdout] | ^^^^^ [INFO] [stdout] 45 | /// Beta 2: [INFO] [stdout] 46 | beta2: f32, [INFO] [stdout] | ^^^^^ [INFO] [stdout] 47 | /// Epsilon [INFO] [stdout] 48 | epsilon: f32, [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] 49 | m_dw: f32, [INFO] [stdout] | ^^^^ [INFO] [stdout] 50 | v_dw: f32, [INFO] [stdout] | ^^^^ [INFO] [stdout] 51 | m_db: f32, [INFO] [stdout] | ^^^^ [INFO] [stdout] 52 | v_db: f32, [INFO] [stdout] | ^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: fields `lr`, `momentum`, and `decay_rate` are never read [INFO] [stdout] --> src/nn/optimizer.rs:97:5 [INFO] [stdout] | [INFO] [stdout] 95 | pub struct SGD { [INFO] [stdout] | --- fields in this struct [INFO] [stdout] 96 | /// Learning rate [INFO] [stdout] 97 | lr: T, [INFO] [stdout] | ^^ [INFO] [stdout] 98 | /// Momentum [INFO] [stdout] 99 | momentum: f32, [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] 100 | /// Decay rate [INFO] [stdout] 101 | decay_rate: f32, [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `swap` is never used [INFO] [stdout] --> src/tensor/mod.rs:33:4 [INFO] [stdout] | [INFO] [stdout] 33 | fn swap(lhs: &mut usize, rhs: &mut usize) { [INFO] [stdout] | ^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: methods `get_sub_tensor`, `sum`, and `prod` are never used [INFO] [stdout] --> src/tensor/mod.rs:686:8 [INFO] [stdout] | [INFO] [stdout] 560 | / impl<'a, T> Tensor<'a, T> [INFO] [stdout] 561 | | where [INFO] [stdout] 562 | | T: TensorElement + Div + Sum, [INFO] [stdout] 563 | | ::Err: Error + 'static, [INFO] [stdout] 564 | | Vec: IntoParallelIterator, [INFO] [stdout] 565 | | Vec<&'a T>: IntoParallelRefIterator<'a>, [INFO] [stdout] | |____________________________________________- methods in this implementation [INFO] [stdout] ... [INFO] [stdout] 686 | fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec { [INFO] [stdout] | ^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 889 | fn sum(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^ [INFO] [stdout] ... [INFO] [stdout] 904 | fn prod(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `ReLU` should have a snake case name [INFO] [stdout] --> src/nn/activation.rs:25:8 [INFO] [stdout] | [INFO] [stdout] 25 | pub fn ReLU<'a, 'b, T>(x: &Tensor<'a, T>) -> Tensor<'b, T> [INFO] [stdout] | ^^^^ help: convert the identifier to snake case: `re_lu` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(non_snake_case)]` (part of `#[warn(nonstandard_style)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `PReLU` should have a snake case name [INFO] [stdout] --> src/nn/activation.rs:52:8 [INFO] [stdout] | [INFO] [stdout] 52 | pub fn PReLU<'a, 'b, T>(x: &Tensor<'a, T>, alpha: T) -> Tensor<'b, T> [INFO] [stdout] | ^^^^^ help: convert the identifier to snake case: `pre_lu` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `Sigmoid` should have a snake case name [INFO] [stdout] --> src/nn/activation.rs:79:8 [INFO] [stdout] | [INFO] [stdout] 79 | pub fn Sigmoid<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T> [INFO] [stdout] | ^^^^^^^ help: convert the identifier to snake case (notice the capitalization): `sigmoid` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `GeLU` should have a snake case name [INFO] [stdout] --> src/nn/activation.rs:108:8 [INFO] [stdout] | [INFO] [stdout] 108 | pub fn GeLU<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T> [INFO] [stdout] | ^^^^ help: convert the identifier to snake case: `ge_lu` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: missing documentation for an associated function [INFO] [stdout] --> src/nn/loss.rs:14:5 [INFO] [stdout] | [INFO] [stdout] 14 | fn loss(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] note: the lint level is defined here [INFO] [stdout] --> src/nn/loss.rs:2:9 [INFO] [stdout] | [INFO] [stdout] 2 | #![warn(missing_docs)] [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: missing documentation for an associated function [INFO] [stdout] --> src/nn/loss.rs:15:5 [INFO] [stdout] | [INFO] [stdout] 15 | fn backward(); [INFO] [stdout] | ^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `BCEntroypyLoss` should have a snake case name [INFO] [stdout] --> src/nn/loss.rs:31:8 [INFO] [stdout] | [INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^^^^^^^^^^ help: convert the identifier to snake case: `bcentroypy_loss` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `CEntroypyLoss` should have a snake case name [INFO] [stdout] --> src/nn/loss.rs:50:8 [INFO] [stdout] | [INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^^^^^^^^^ help: convert the identifier to snake case: `centroypy_loss` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `L1Loss` should have a snake case name [INFO] [stdout] --> src/nn/loss.rs:73:8 [INFO] [stdout] | [INFO] [stdout] 73 | pub fn L1Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^^ help: convert the identifier to snake case: `l1_loss` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `L2Loss` should have a snake case name [INFO] [stdout] --> src/nn/loss.rs:96:8 [INFO] [stdout] | [INFO] [stdout] 96 | pub fn L2Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^^ help: convert the identifier to snake case: `l2_loss` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `MaxPool` should have a snake case name [INFO] [stdout] --> src/nn/pooling.rs:23:8 [INFO] [stdout] | [INFO] [stdout] 23 | pub fn MaxPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T> [INFO] [stdout] | ^^^^^^^ help: convert the identifier to snake case: `max_pool` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `MinPool` should have a snake case name [INFO] [stdout] --> src/nn/pooling.rs:62:8 [INFO] [stdout] | [INFO] [stdout] 62 | pub fn MinPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T> [INFO] [stdout] | ^^^^^^^ help: convert the identifier to snake case: `min_pool` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `AvgPool` should have a snake case name [INFO] [stdout] --> src/nn/pooling.rs:93:8 [INFO] [stdout] | [INFO] [stdout] 93 | pub fn AvgPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T> [INFO] [stdout] | ^^^^^^^ help: convert the identifier to snake case: `avg_pool` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: missing documentation for a struct [INFO] [stdout] --> src/nn/transform.rs:4:1 [INFO] [stdout] | [INFO] [stdout] 4 | pub struct Transform; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] note: the lint level is defined here [INFO] [stdout] --> src/nn/transform.rs:2:9 [INFO] [stdout] | [INFO] [stdout] 2 | #![warn(missing_docs)] [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: missing documentation for a struct [INFO] [stdout] --> src/tensor/mod.rs:145:1 [INFO] [stdout] | [INFO] [stdout] 145 | pub struct Tensor<'a, T> [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] note: the lint level is defined here [INFO] [stdout] --> src/lib.rs:1:9 [INFO] [stdout] | [INFO] [stdout] 1 | #![warn(missing_docs)] [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] Finished `dev` profile [unoptimized + debuginfo] target(s) in 13.15s [INFO] running `Command { std: "docker" "inspect" "24cd81002956ca7d25a15907c53115dea16a1ba750654bc64b0367ed2ffbb797", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "24cd81002956ca7d25a15907c53115dea16a1ba750654bc64b0367ed2ffbb797", kill_on_drop: false }` [INFO] [stdout] 24cd81002956ca7d25a15907c53115dea16a1ba750654bc64b0367ed2ffbb797 [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:4848fb76d95f26979359cc7e45710b1dbc8f3acb7aeedee7c460d7702230f228" "/opt/rustwide/cargo-home/bin/cargo" "+c2e32f1c9652b13ed99608599c1e855462f421f3" "test" "--frozen" "--no-run" "--message-format=json", kill_on_drop: false }` [INFO] [stdout] 635b5c25ed06deac579c9df45acee66562d04182cf8a4d431d41ad2f84ceb3dc [INFO] running `Command { std: "docker" "start" "-a" "635b5c25ed06deac579c9df45acee66562d04182cf8a4d431d41ad2f84ceb3dc", kill_on_drop: false }` [INFO] [stderr] Compiling libc v0.2.146 [INFO] [stderr] Compiling serde v1.0.164 [INFO] [stderr] Compiling bitflags v1.3.2 [INFO] [stderr] Compiling io-lifetimes v1.0.11 [INFO] [stderr] Compiling rustix v0.37.19 [INFO] [stderr] Compiling serde_json v1.0.96 [INFO] [stderr] Compiling linux-raw-sys v0.3.8 [INFO] [stderr] Compiling anstyle v1.0.0 [INFO] [stderr] Compiling clap_lex v0.5.0 [INFO] [stderr] Compiling half v1.8.2 [INFO] [stderr] Compiling ciborium-io v0.2.1 [INFO] [stderr] Compiling plotters-backend v0.3.4 [INFO] [stderr] Compiling itoa v1.0.6 [INFO] [stderr] Compiling ryu v1.0.13 [INFO] [stderr] Compiling same-file v1.0.6 [INFO] [stderr] Compiling cast v0.3.0 [INFO] [stderr] Compiling regex-syntax v0.7.2 [INFO] [stderr] Compiling ciborium-ll v0.2.1 [INFO] [stderr] Compiling once_cell v1.18.0 [INFO] [stderr] Compiling anes v0.1.6 [INFO] [stderr] Compiling plotters-svg v0.3.3 [INFO] [stderr] Compiling clap_builder v4.3.3 [INFO] [stderr] Compiling walkdir v2.3.3 [INFO] [stderr] Compiling oorandom v11.1.3 [INFO] [stderr] Compiling criterion-plot v0.5.0 [INFO] [stderr] Compiling plotters v0.3.4 [INFO] [stderr] Compiling num_cpus v1.15.0 [INFO] [stderr] Compiling getrandom v0.2.10 [INFO] [stderr] Compiling rand_core v0.6.4 [INFO] [stderr] Compiling rayon-core v1.11.0 [INFO] [stderr] Compiling regex v1.8.4 [INFO] [stderr] Compiling rand_chacha v0.3.1 [INFO] [stderr] Compiling rand v0.8.5 [INFO] [stderr] Compiling rayon v1.7.0 [INFO] [stderr] Compiling clap v4.3.3 [INFO] [stderr] Compiling is-terminal v0.4.7 [INFO] [stderr] Compiling ciborium v0.2.1 [INFO] [stderr] Compiling kaffe v0.2.0 (/opt/rustwide/workdir) [INFO] [stdout] warning: unused import: `marker::PhantomData` [INFO] [stdout] --> src/nn/optimizer.rs:4:25 [INFO] [stdout] | [INFO] [stdout] 4 | use std::{error::Error, marker::PhantomData, str::FromStr}; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `rayon::prelude::*` [INFO] [stdout] --> src/nn/optimizer.rs:7:5 [INFO] [stdout] | [INFO] [stdout] 7 | use rayon::prelude::*; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] Compiling tinytemplate v1.2.1 [INFO] [stdout] warning: unused variable: `idx` [INFO] [stdout] --> src/nn/dataset.rs:57:26 [INFO] [stdout] | [INFO] [stdout] 57 | fn get<'a, T>(&self, idx: usize) -> (Tensor<'a, T>, String) [INFO] [stdout] | ^^^ help: if this is intentional, prefix it with an underscore: `_idx` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_variables)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `y` [INFO] [stdout] --> src/nn/loss.rs:31:30 [INFO] [stdout] | [INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_y` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `y_hat` [INFO] [stdout] --> src/nn/loss.rs:31:49 [INFO] [stdout] | [INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `y` [INFO] [stdout] --> src/nn/loss.rs:50:29 [INFO] [stdout] | [INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_y` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `y_hat` [INFO] [stdout] --> src/nn/loss.rs:50:48 [INFO] [stdout] | [INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `cost` [INFO] [stdout] --> src/nn/optimizer.rs:75:31 [INFO] [stdout] | [INFO] [stdout] 75 | fn minimize(&mut self, cost: F, vars: &mut Vec) [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_cost` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `vars` [INFO] [stdout] --> src/nn/optimizer.rs:75:40 [INFO] [stdout] | [INFO] [stdout] 75 | fn minimize(&mut self, cost: F, vars: &mut Vec) [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_vars` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `cost` [INFO] [stdout] --> src/nn/optimizer.rs:117:31 [INFO] [stdout] | [INFO] [stdout] 117 | fn minimize(&mut self, cost: F, vars: &mut Vec) [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_cost` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `vars` [INFO] [stdout] --> src/nn/optimizer.rs:117:40 [INFO] [stdout] | [INFO] [stdout] 117 | fn minimize(&mut self, cost: F, vars: &mut Vec) [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_vars` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `tensor` [INFO] [stdout] --> src/nn/mod.rs:24:5 [INFO] [stdout] | [INFO] [stdout] 24 | tensor: Tensor<'a, T>, [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `kernel` [INFO] [stdout] --> src/nn/mod.rs:25:5 [INFO] [stdout] | [INFO] [stdout] 25 | kernel: Tensor<'a, T>, [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_kernel` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `stride` [INFO] [stdout] --> src/nn/mod.rs:26:5 [INFO] [stdout] | [INFO] [stdout] 26 | stride: usize, [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_stride` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `padding` [INFO] [stdout] --> src/nn/mod.rs:27:5 [INFO] [stdout] | [INFO] [stdout] 27 | padding: usize, [INFO] [stdout] | ^^^^^^^ help: if this is intentional, prefix it with an underscore: `_padding` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `f` [INFO] [stdout] --> src/tensor/mod.rs:257:19 [INFO] [stdout] | [INFO] [stdout] 257 | fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_f` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `start_idx` [INFO] [stdout] --> src/tensor/mod.rs:686:30 [INFO] [stdout] | [INFO] [stdout] 686 | fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec { [INFO] [stdout] | ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_start_idx` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `size` [INFO] [stdout] --> src/tensor/mod.rs:686:48 [INFO] [stdout] | [INFO] [stdout] 686 | fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec { [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_size` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `rowcol` [INFO] [stdout] --> src/tensor/mod.rs:889:19 [INFO] [stdout] | [INFO] [stdout] 889 | fn sum(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `dimension` [INFO] [stdout] --> src/tensor/mod.rs:889:34 [INFO] [stdout] | [INFO] [stdout] 889 | fn sum(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `rowcol` [INFO] [stdout] --> src/tensor/mod.rs:904:20 [INFO] [stdout] | [INFO] [stdout] 904 | fn prod(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `dimension` [INFO] [stdout] --> src/tensor/mod.rs:904:35 [INFO] [stdout] | [INFO] [stdout] 904 | fn prod(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `layers` is never read [INFO] [stdout] --> src/nn/mod.rs:39:5 [INFO] [stdout] | [INFO] [stdout] 37 | pub struct Net { [INFO] [stdout] | --- field in this struct [INFO] [stdout] 38 | /// Represents all layers in the network [INFO] [stdout] 39 | layers: Vec, [INFO] [stdout] | ^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: fields `data`, `batch_size`, and `shuffle` are never read [INFO] [stdout] --> src/nn/dataloader.rs:10:5 [INFO] [stdout] | [INFO] [stdout] 8 | pub struct DataLoader { [INFO] [stdout] | ---------- fields in this struct [INFO] [stdout] 9 | /// Dataset [INFO] [stdout] 10 | data: D, [INFO] [stdout] | ^^^^ [INFO] [stdout] 11 | /// Batch size [INFO] [stdout] 12 | batch_size: usize, [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] 13 | /// Whether or not to shuffle data [INFO] [stdout] 14 | shuffle: bool, [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `DataLoader` has a derived impl for the trait `Clone`, but this is intentionally ignored during dead code analysis [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: associated items `new` and `iter` are never used [INFO] [stdout] --> src/nn/dataloader.rs:41:8 [INFO] [stdout] | [INFO] [stdout] 33 | impl DataLoader { [INFO] [stdout] | -------------------------------------- associated items in this implementation [INFO] [stdout] ... [INFO] [stdout] 41 | fn new(data: D, batch_size: usize, shuffle: bool) -> DataLoader { [INFO] [stdout] | ^^^ [INFO] [stdout] ... [INFO] [stdout] 56 | fn iter(&mut self) -> DataLoader { [INFO] [stdout] | ^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: fields `img_dir`, `transform`, and `target_transform` are never read [INFO] [stdout] --> src/nn/dataset.rs:33:5 [INFO] [stdout] | [INFO] [stdout] 31 | pub struct CustomImageDataset { [INFO] [stdout] | ------------------ fields in this struct [INFO] [stdout] 32 | img_labels: Vec, [INFO] [stdout] 33 | img_dir: &'static str, [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] 34 | transform: Option, [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] 35 | target_transform: Option, [INFO] [stdout] | ^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: multiple fields are never read [INFO] [stdout] --> src/nn/optimizer.rs:38:5 [INFO] [stdout] | [INFO] [stdout] 36 | pub struct Adam { [INFO] [stdout] | ---- fields in this struct [INFO] [stdout] 37 | /// Learning rate [INFO] [stdout] 38 | lr: T, [INFO] [stdout] | ^^ [INFO] [stdout] 39 | /// Momentum [INFO] [stdout] 40 | momentum: f32, [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] 41 | /// Decay rate [INFO] [stdout] 42 | decay_rate: f32, [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] 43 | /// Beta 1 [INFO] [stdout] 44 | beta1: f32, [INFO] [stdout] | ^^^^^ [INFO] [stdout] 45 | /// Beta 2: [INFO] [stdout] 46 | beta2: f32, [INFO] [stdout] | ^^^^^ [INFO] [stdout] 47 | /// Epsilon [INFO] [stdout] 48 | epsilon: f32, [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] 49 | m_dw: f32, [INFO] [stdout] | ^^^^ [INFO] [stdout] 50 | v_dw: f32, [INFO] [stdout] | ^^^^ [INFO] [stdout] 51 | m_db: f32, [INFO] [stdout] | ^^^^ [INFO] [stdout] 52 | v_db: f32, [INFO] [stdout] | ^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: fields `lr`, `momentum`, and `decay_rate` are never read [INFO] [stdout] --> src/nn/optimizer.rs:97:5 [INFO] [stdout] | [INFO] [stdout] 95 | pub struct SGD { [INFO] [stdout] | --- fields in this struct [INFO] [stdout] 96 | /// Learning rate [INFO] [stdout] 97 | lr: T, [INFO] [stdout] | ^^ [INFO] [stdout] 98 | /// Momentum [INFO] [stdout] 99 | momentum: f32, [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] 100 | /// Decay rate [INFO] [stdout] 101 | decay_rate: f32, [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `swap` is never used [INFO] [stdout] --> src/tensor/mod.rs:33:4 [INFO] [stdout] | [INFO] [stdout] 33 | fn swap(lhs: &mut usize, rhs: &mut usize) { [INFO] [stdout] | ^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: methods `get_sub_tensor`, `sum`, and `prod` are never used [INFO] [stdout] --> src/tensor/mod.rs:686:8 [INFO] [stdout] | [INFO] [stdout] 560 | / impl<'a, T> Tensor<'a, T> [INFO] [stdout] 561 | | where [INFO] [stdout] 562 | | T: TensorElement + Div + Sum, [INFO] [stdout] 563 | | ::Err: Error + 'static, [INFO] [stdout] 564 | | Vec: IntoParallelIterator, [INFO] [stdout] 565 | | Vec<&'a T>: IntoParallelRefIterator<'a>, [INFO] [stdout] | |____________________________________________- methods in this implementation [INFO] [stdout] ... [INFO] [stdout] 686 | fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec { [INFO] [stdout] | ^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 889 | fn sum(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^ [INFO] [stdout] ... [INFO] [stdout] 904 | fn prod(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `ReLU` should have a snake case name [INFO] [stdout] --> src/nn/activation.rs:25:8 [INFO] [stdout] | [INFO] [stdout] 25 | pub fn ReLU<'a, 'b, T>(x: &Tensor<'a, T>) -> Tensor<'b, T> [INFO] [stdout] | ^^^^ help: convert the identifier to snake case: `re_lu` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(non_snake_case)]` (part of `#[warn(nonstandard_style)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `PReLU` should have a snake case name [INFO] [stdout] --> src/nn/activation.rs:52:8 [INFO] [stdout] | [INFO] [stdout] 52 | pub fn PReLU<'a, 'b, T>(x: &Tensor<'a, T>, alpha: T) -> Tensor<'b, T> [INFO] [stdout] | ^^^^^ help: convert the identifier to snake case: `pre_lu` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `Sigmoid` should have a snake case name [INFO] [stdout] --> src/nn/activation.rs:79:8 [INFO] [stdout] | [INFO] [stdout] 79 | pub fn Sigmoid<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T> [INFO] [stdout] | ^^^^^^^ help: convert the identifier to snake case (notice the capitalization): `sigmoid` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `GeLU` should have a snake case name [INFO] [stdout] --> src/nn/activation.rs:108:8 [INFO] [stdout] | [INFO] [stdout] 108 | pub fn GeLU<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T> [INFO] [stdout] | ^^^^ help: convert the identifier to snake case: `ge_lu` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: missing documentation for an associated function [INFO] [stdout] --> src/nn/loss.rs:14:5 [INFO] [stdout] | [INFO] [stdout] 14 | fn loss(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] note: the lint level is defined here [INFO] [stdout] --> src/nn/loss.rs:2:9 [INFO] [stdout] | [INFO] [stdout] 2 | #![warn(missing_docs)] [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: missing documentation for an associated function [INFO] [stdout] --> src/nn/loss.rs:15:5 [INFO] [stdout] | [INFO] [stdout] 15 | fn backward(); [INFO] [stdout] | ^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `BCEntroypyLoss` should have a snake case name [INFO] [stdout] --> src/nn/loss.rs:31:8 [INFO] [stdout] | [INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^^^^^^^^^^ help: convert the identifier to snake case: `bcentroypy_loss` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `CEntroypyLoss` should have a snake case name [INFO] [stdout] --> src/nn/loss.rs:50:8 [INFO] [stdout] | [INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^^^^^^^^^ help: convert the identifier to snake case: `centroypy_loss` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `L1Loss` should have a snake case name [INFO] [stdout] --> src/nn/loss.rs:73:8 [INFO] [stdout] | [INFO] [stdout] 73 | pub fn L1Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^^ help: convert the identifier to snake case: `l1_loss` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `L2Loss` should have a snake case name [INFO] [stdout] --> src/nn/loss.rs:96:8 [INFO] [stdout] | [INFO] [stdout] 96 | pub fn L2Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^^ help: convert the identifier to snake case: `l2_loss` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `MaxPool` should have a snake case name [INFO] [stdout] --> src/nn/pooling.rs:23:8 [INFO] [stdout] | [INFO] [stdout] 23 | pub fn MaxPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T> [INFO] [stdout] | ^^^^^^^ help: convert the identifier to snake case: `max_pool` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `MinPool` should have a snake case name [INFO] [stdout] --> src/nn/pooling.rs:62:8 [INFO] [stdout] | [INFO] [stdout] 62 | pub fn MinPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T> [INFO] [stdout] | ^^^^^^^ help: convert the identifier to snake case: `min_pool` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `AvgPool` should have a snake case name [INFO] [stdout] --> src/nn/pooling.rs:93:8 [INFO] [stdout] | [INFO] [stdout] 93 | pub fn AvgPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T> [INFO] [stdout] | ^^^^^^^ help: convert the identifier to snake case: `avg_pool` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: missing documentation for a struct [INFO] [stdout] --> src/nn/transform.rs:4:1 [INFO] [stdout] | [INFO] [stdout] 4 | pub struct Transform; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] note: the lint level is defined here [INFO] [stdout] --> src/nn/transform.rs:2:9 [INFO] [stdout] | [INFO] [stdout] 2 | #![warn(missing_docs)] [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: missing documentation for a struct [INFO] [stdout] --> src/tensor/mod.rs:145:1 [INFO] [stdout] | [INFO] [stdout] 145 | pub struct Tensor<'a, T> [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] note: the lint level is defined here [INFO] [stdout] --> src/lib.rs:1:9 [INFO] [stdout] | [INFO] [stdout] 1 | #![warn(missing_docs)] [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] Compiling criterion v0.5.1 [INFO] [stdout] warning: unused import: `marker::PhantomData` [INFO] [stdout] --> src/nn/optimizer.rs:4:25 [INFO] [stdout] | [INFO] [stdout] 4 | use std::{error::Error, marker::PhantomData, str::FromStr}; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `rayon::prelude::*` [INFO] [stdout] --> src/nn/optimizer.rs:7:5 [INFO] [stdout] | [INFO] [stdout] 7 | use rayon::prelude::*; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `idx` [INFO] [stdout] --> src/nn/dataset.rs:57:26 [INFO] [stdout] | [INFO] [stdout] 57 | fn get<'a, T>(&self, idx: usize) -> (Tensor<'a, T>, String) [INFO] [stdout] | ^^^ help: if this is intentional, prefix it with an underscore: `_idx` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_variables)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `y` [INFO] [stdout] --> src/nn/loss.rs:31:30 [INFO] [stdout] | [INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_y` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `y_hat` [INFO] [stdout] --> src/nn/loss.rs:31:49 [INFO] [stdout] | [INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `y` [INFO] [stdout] --> src/nn/loss.rs:50:29 [INFO] [stdout] | [INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_y` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `y_hat` [INFO] [stdout] --> src/nn/loss.rs:50:48 [INFO] [stdout] | [INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `cost` [INFO] [stdout] --> src/nn/optimizer.rs:75:31 [INFO] [stdout] | [INFO] [stdout] 75 | fn minimize(&mut self, cost: F, vars: &mut Vec) [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_cost` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `vars` [INFO] [stdout] --> src/nn/optimizer.rs:75:40 [INFO] [stdout] | [INFO] [stdout] 75 | fn minimize(&mut self, cost: F, vars: &mut Vec) [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_vars` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `cost` [INFO] [stdout] --> src/nn/optimizer.rs:117:31 [INFO] [stdout] | [INFO] [stdout] 117 | fn minimize(&mut self, cost: F, vars: &mut Vec) [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_cost` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `vars` [INFO] [stdout] --> src/nn/optimizer.rs:117:40 [INFO] [stdout] | [INFO] [stdout] 117 | fn minimize(&mut self, cost: F, vars: &mut Vec) [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_vars` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `tensor` [INFO] [stdout] --> src/nn/mod.rs:24:5 [INFO] [stdout] | [INFO] [stdout] 24 | tensor: Tensor<'a, T>, [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `kernel` [INFO] [stdout] --> src/nn/mod.rs:25:5 [INFO] [stdout] | [INFO] [stdout] 25 | kernel: Tensor<'a, T>, [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_kernel` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `stride` [INFO] [stdout] --> src/nn/mod.rs:26:5 [INFO] [stdout] | [INFO] [stdout] 26 | stride: usize, [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_stride` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `padding` [INFO] [stdout] --> src/nn/mod.rs:27:5 [INFO] [stdout] | [INFO] [stdout] 27 | padding: usize, [INFO] [stdout] | ^^^^^^^ help: if this is intentional, prefix it with an underscore: `_padding` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `f` [INFO] [stdout] --> src/tensor/mod.rs:257:19 [INFO] [stdout] | [INFO] [stdout] 257 | fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_f` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `start_idx` [INFO] [stdout] --> src/tensor/mod.rs:686:30 [INFO] [stdout] | [INFO] [stdout] 686 | fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec { [INFO] [stdout] | ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_start_idx` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `size` [INFO] [stdout] --> src/tensor/mod.rs:686:48 [INFO] [stdout] | [INFO] [stdout] 686 | fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec { [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_size` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `rowcol` [INFO] [stdout] --> src/tensor/mod.rs:889:19 [INFO] [stdout] | [INFO] [stdout] 889 | fn sum(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `dimension` [INFO] [stdout] --> src/tensor/mod.rs:889:34 [INFO] [stdout] | [INFO] [stdout] 889 | fn sum(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `rowcol` [INFO] [stdout] --> src/tensor/mod.rs:904:20 [INFO] [stdout] | [INFO] [stdout] 904 | fn prod(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `dimension` [INFO] [stdout] --> src/tensor/mod.rs:904:35 [INFO] [stdout] | [INFO] [stdout] 904 | fn prod(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `layers` is never read [INFO] [stdout] --> src/nn/mod.rs:39:5 [INFO] [stdout] | [INFO] [stdout] 37 | pub struct Net { [INFO] [stdout] | --- field in this struct [INFO] [stdout] 38 | /// Represents all layers in the network [INFO] [stdout] 39 | layers: Vec, [INFO] [stdout] | ^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: fields `data`, `batch_size`, and `shuffle` are never read [INFO] [stdout] --> src/nn/dataloader.rs:10:5 [INFO] [stdout] | [INFO] [stdout] 8 | pub struct DataLoader { [INFO] [stdout] | ---------- fields in this struct [INFO] [stdout] 9 | /// Dataset [INFO] [stdout] 10 | data: D, [INFO] [stdout] | ^^^^ [INFO] [stdout] 11 | /// Batch size [INFO] [stdout] 12 | batch_size: usize, [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] 13 | /// Whether or not to shuffle data [INFO] [stdout] 14 | shuffle: bool, [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `DataLoader` has a derived impl for the trait `Clone`, but this is intentionally ignored during dead code analysis [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: associated items `new` and `iter` are never used [INFO] [stdout] --> src/nn/dataloader.rs:41:8 [INFO] [stdout] | [INFO] [stdout] 33 | impl DataLoader { [INFO] [stdout] | -------------------------------------- associated items in this implementation [INFO] [stdout] ... [INFO] [stdout] 41 | fn new(data: D, batch_size: usize, shuffle: bool) -> DataLoader { [INFO] [stdout] | ^^^ [INFO] [stdout] ... [INFO] [stdout] 56 | fn iter(&mut self) -> DataLoader { [INFO] [stdout] | ^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: fields `img_dir`, `transform`, and `target_transform` are never read [INFO] [stdout] --> src/nn/dataset.rs:33:5 [INFO] [stdout] | [INFO] [stdout] 31 | pub struct CustomImageDataset { [INFO] [stdout] | ------------------ fields in this struct [INFO] [stdout] 32 | img_labels: Vec, [INFO] [stdout] 33 | img_dir: &'static str, [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] 34 | transform: Option, [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] 35 | target_transform: Option, [INFO] [stdout] | ^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: multiple fields are never read [INFO] [stdout] --> src/nn/optimizer.rs:38:5 [INFO] [stdout] | [INFO] [stdout] 36 | pub struct Adam { [INFO] [stdout] | ---- fields in this struct [INFO] [stdout] 37 | /// Learning rate [INFO] [stdout] 38 | lr: T, [INFO] [stdout] | ^^ [INFO] [stdout] 39 | /// Momentum [INFO] [stdout] 40 | momentum: f32, [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] 41 | /// Decay rate [INFO] [stdout] 42 | decay_rate: f32, [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] 43 | /// Beta 1 [INFO] [stdout] 44 | beta1: f32, [INFO] [stdout] | ^^^^^ [INFO] [stdout] 45 | /// Beta 2: [INFO] [stdout] 46 | beta2: f32, [INFO] [stdout] | ^^^^^ [INFO] [stdout] 47 | /// Epsilon [INFO] [stdout] 48 | epsilon: f32, [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] 49 | m_dw: f32, [INFO] [stdout] | ^^^^ [INFO] [stdout] 50 | v_dw: f32, [INFO] [stdout] | ^^^^ [INFO] [stdout] 51 | m_db: f32, [INFO] [stdout] | ^^^^ [INFO] [stdout] 52 | v_db: f32, [INFO] [stdout] | ^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: fields `lr`, `momentum`, and `decay_rate` are never read [INFO] [stdout] --> src/nn/optimizer.rs:97:5 [INFO] [stdout] | [INFO] [stdout] 95 | pub struct SGD { [INFO] [stdout] | --- fields in this struct [INFO] [stdout] 96 | /// Learning rate [INFO] [stdout] 97 | lr: T, [INFO] [stdout] | ^^ [INFO] [stdout] 98 | /// Momentum [INFO] [stdout] 99 | momentum: f32, [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] 100 | /// Decay rate [INFO] [stdout] 101 | decay_rate: f32, [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `swap` is never used [INFO] [stdout] --> src/tensor/mod.rs:33:4 [INFO] [stdout] | [INFO] [stdout] 33 | fn swap(lhs: &mut usize, rhs: &mut usize) { [INFO] [stdout] | ^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: methods `get_sub_tensor`, `sum`, and `prod` are never used [INFO] [stdout] --> src/tensor/mod.rs:686:8 [INFO] [stdout] | [INFO] [stdout] 560 | / impl<'a, T> Tensor<'a, T> [INFO] [stdout] 561 | | where [INFO] [stdout] 562 | | T: TensorElement + Div + Sum, [INFO] [stdout] 563 | | ::Err: Error + 'static, [INFO] [stdout] 564 | | Vec: IntoParallelIterator, [INFO] [stdout] 565 | | Vec<&'a T>: IntoParallelRefIterator<'a>, [INFO] [stdout] | |____________________________________________- methods in this implementation [INFO] [stdout] ... [INFO] [stdout] 686 | fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec { [INFO] [stdout] | ^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 889 | fn sum(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^ [INFO] [stdout] ... [INFO] [stdout] 904 | fn prod(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `ReLU` should have a snake case name [INFO] [stdout] --> src/nn/activation.rs:25:8 [INFO] [stdout] | [INFO] [stdout] 25 | pub fn ReLU<'a, 'b, T>(x: &Tensor<'a, T>) -> Tensor<'b, T> [INFO] [stdout] | ^^^^ help: convert the identifier to snake case: `re_lu` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(non_snake_case)]` (part of `#[warn(nonstandard_style)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `PReLU` should have a snake case name [INFO] [stdout] --> src/nn/activation.rs:52:8 [INFO] [stdout] | [INFO] [stdout] 52 | pub fn PReLU<'a, 'b, T>(x: &Tensor<'a, T>, alpha: T) -> Tensor<'b, T> [INFO] [stdout] | ^^^^^ help: convert the identifier to snake case: `pre_lu` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `Sigmoid` should have a snake case name [INFO] [stdout] --> src/nn/activation.rs:79:8 [INFO] [stdout] | [INFO] [stdout] 79 | pub fn Sigmoid<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T> [INFO] [stdout] | ^^^^^^^ help: convert the identifier to snake case (notice the capitalization): `sigmoid` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `GeLU` should have a snake case name [INFO] [stdout] --> src/nn/activation.rs:108:8 [INFO] [stdout] | [INFO] [stdout] 108 | pub fn GeLU<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T> [INFO] [stdout] | ^^^^ help: convert the identifier to snake case: `ge_lu` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: missing documentation for an associated function [INFO] [stdout] --> src/nn/loss.rs:14:5 [INFO] [stdout] | [INFO] [stdout] 14 | fn loss(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] note: the lint level is defined here [INFO] [stdout] --> src/nn/loss.rs:2:9 [INFO] [stdout] | [INFO] [stdout] 2 | #![warn(missing_docs)] [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: missing documentation for an associated function [INFO] [stdout] --> src/nn/loss.rs:15:5 [INFO] [stdout] | [INFO] [stdout] 15 | fn backward(); [INFO] [stdout] | ^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `BCEntroypyLoss` should have a snake case name [INFO] [stdout] --> src/nn/loss.rs:31:8 [INFO] [stdout] | [INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^^^^^^^^^^ help: convert the identifier to snake case: `bcentroypy_loss` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `CEntroypyLoss` should have a snake case name [INFO] [stdout] --> src/nn/loss.rs:50:8 [INFO] [stdout] | [INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^^^^^^^^^ help: convert the identifier to snake case: `centroypy_loss` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `L1Loss` should have a snake case name [INFO] [stdout] --> src/nn/loss.rs:73:8 [INFO] [stdout] | [INFO] [stdout] 73 | pub fn L1Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^^ help: convert the identifier to snake case: `l1_loss` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `L2Loss` should have a snake case name [INFO] [stdout] --> src/nn/loss.rs:96:8 [INFO] [stdout] | [INFO] [stdout] 96 | pub fn L2Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^^ help: convert the identifier to snake case: `l2_loss` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `MaxPool` should have a snake case name [INFO] [stdout] --> src/nn/pooling.rs:23:8 [INFO] [stdout] | [INFO] [stdout] 23 | pub fn MaxPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T> [INFO] [stdout] | ^^^^^^^ help: convert the identifier to snake case: `max_pool` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `MinPool` should have a snake case name [INFO] [stdout] --> src/nn/pooling.rs:62:8 [INFO] [stdout] | [INFO] [stdout] 62 | pub fn MinPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T> [INFO] [stdout] | ^^^^^^^ help: convert the identifier to snake case: `min_pool` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `AvgPool` should have a snake case name [INFO] [stdout] --> src/nn/pooling.rs:93:8 [INFO] [stdout] | [INFO] [stdout] 93 | pub fn AvgPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T> [INFO] [stdout] | ^^^^^^^ help: convert the identifier to snake case: `avg_pool` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: missing documentation for a struct [INFO] [stdout] --> src/nn/transform.rs:4:1 [INFO] [stdout] | [INFO] [stdout] 4 | pub struct Transform; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] note: the lint level is defined here [INFO] [stdout] --> src/nn/transform.rs:2:9 [INFO] [stdout] | [INFO] [stdout] 2 | #![warn(missing_docs)] [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: missing documentation for a struct [INFO] [stdout] --> src/tensor/mod.rs:145:1 [INFO] [stdout] | [INFO] [stdout] 145 | pub struct Tensor<'a, T> [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] note: the lint level is defined here [INFO] [stdout] --> src/lib.rs:1:9 [INFO] [stdout] | [INFO] [stdout] 1 | #![warn(missing_docs)] [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] Finished `test` profile [unoptimized + debuginfo] target(s) in 21.04s [INFO] running `Command { std: "docker" "inspect" "635b5c25ed06deac579c9df45acee66562d04182cf8a4d431d41ad2f84ceb3dc", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "635b5c25ed06deac579c9df45acee66562d04182cf8a4d431d41ad2f84ceb3dc", kill_on_drop: false }` [INFO] [stdout] 635b5c25ed06deac579c9df45acee66562d04182cf8a4d431d41ad2f84ceb3dc [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:4848fb76d95f26979359cc7e45710b1dbc8f3acb7aeedee7c460d7702230f228" "/opt/rustwide/cargo-home/bin/cargo" "+c2e32f1c9652b13ed99608599c1e855462f421f3" "test" "--frozen", kill_on_drop: false }` [INFO] [stdout] 4b0eda8e16ea00dc4c4f01466c4c93aae3c72defb22f1eace42fe9bbe2ff9fd5 [INFO] running `Command { std: "docker" "start" "-a" "4b0eda8e16ea00dc4c4f01466c4c93aae3c72defb22f1eace42fe9bbe2ff9fd5", kill_on_drop: false }` [INFO] [stderr] warning: unused import: `marker::PhantomData` [INFO] [stderr] --> src/nn/optimizer.rs:4:25 [INFO] [stderr] | [INFO] [stderr] 4 | use std::{error::Error, marker::PhantomData, str::FromStr}; [INFO] [stderr] | ^^^^^^^^^^^^^^^^^^^ [INFO] [stderr] | [INFO] [stderr] = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default [INFO] [stderr] [INFO] [stderr] warning: unused import: `rayon::prelude::*` [INFO] [stderr] --> src/nn/optimizer.rs:7:5 [INFO] [stderr] | [INFO] [stderr] 7 | use rayon::prelude::*; [INFO] [stderr] | ^^^^^^^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused variable: `idx` [INFO] [stderr] --> src/nn/dataset.rs:57:26 [INFO] [stderr] | [INFO] [stderr] 57 | fn get<'a, T>(&self, idx: usize) -> (Tensor<'a, T>, String) [INFO] [stderr] | ^^^ help: if this is intentional, prefix it with an underscore: `_idx` [INFO] [stderr] | [INFO] [stderr] = note: `#[warn(unused_variables)]` (part of `#[warn(unused)]`) on by default [INFO] [stderr] [INFO] [stderr] warning: unused variable: `y` [INFO] [stderr] --> src/nn/loss.rs:31:30 [INFO] [stderr] | [INFO] [stderr] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stderr] | ^ help: if this is intentional, prefix it with an underscore: `_y` [INFO] [stderr] [INFO] [stderr] warning: unused variable: `y_hat` [INFO] [stderr] --> src/nn/loss.rs:31:49 [INFO] [stderr] | [INFO] [stderr] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stderr] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat` [INFO] [stderr] [INFO] [stderr] warning: unused variable: `y` [INFO] [stderr] --> src/nn/loss.rs:50:29 [INFO] [stderr] | [INFO] [stderr] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stderr] | ^ help: if this is intentional, prefix it with an underscore: `_y` [INFO] [stderr] [INFO] [stderr] warning: unused variable: `y_hat` [INFO] [stderr] --> src/nn/loss.rs:50:48 [INFO] [stderr] | [INFO] [stderr] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stderr] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat` [INFO] [stderr] [INFO] [stderr] warning: unused variable: `cost` [INFO] [stderr] --> src/nn/optimizer.rs:75:31 [INFO] [stderr] | [INFO] [stderr] 75 | fn minimize(&mut self, cost: F, vars: &mut Vec) [INFO] [stderr] | ^^^^ help: if this is intentional, prefix it with an underscore: `_cost` [INFO] [stderr] [INFO] [stderr] warning: unused variable: `vars` [INFO] [stderr] --> src/nn/optimizer.rs:75:40 [INFO] [stderr] | [INFO] [stderr] 75 | fn minimize(&mut self, cost: F, vars: &mut Vec) [INFO] [stderr] | ^^^^ help: if this is intentional, prefix it with an underscore: `_vars` [INFO] [stderr] [INFO] [stderr] warning: unused variable: `cost` [INFO] [stderr] --> src/nn/optimizer.rs:117:31 [INFO] [stderr] | [INFO] [stderr] 117 | fn minimize(&mut self, cost: F, vars: &mut Vec) [INFO] [stderr] | ^^^^ help: if this is intentional, prefix it with an underscore: `_cost` [INFO] [stderr] [INFO] [stderr] warning: unused variable: `vars` [INFO] [stderr] --> src/nn/optimizer.rs:117:40 [INFO] [stderr] | [INFO] [stderr] 117 | fn minimize(&mut self, cost: F, vars: &mut Vec) [INFO] [stderr] | ^^^^ help: if this is intentional, prefix it with an underscore: `_vars` [INFO] [stderr] [INFO] [stderr] warning: unused variable: `tensor` [INFO] [stderr] --> src/nn/mod.rs:24:5 [INFO] [stderr] | [INFO] [stderr] 24 | tensor: Tensor<'a, T>, [INFO] [stderr] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_tensor` [INFO] [stderr] [INFO] [stderr] warning: unused variable: `kernel` [INFO] [stderr] --> src/nn/mod.rs:25:5 [INFO] [stderr] | [INFO] [stderr] 25 | kernel: Tensor<'a, T>, [INFO] [stderr] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_kernel` [INFO] [stderr] [INFO] [stderr] warning: unused variable: `stride` [INFO] [stderr] --> src/nn/mod.rs:26:5 [INFO] [stderr] | [INFO] [stderr] 26 | stride: usize, [INFO] [stderr] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_stride` [INFO] [stderr] [INFO] [stderr] warning: unused variable: `padding` [INFO] [stderr] --> src/nn/mod.rs:27:5 [INFO] [stderr] | [INFO] [stderr] 27 | padding: usize, [INFO] [stderr] | ^^^^^^^ help: if this is intentional, prefix it with an underscore: `_padding` [INFO] [stderr] [INFO] [stderr] warning: unused variable: `f` [INFO] [stderr] --> src/tensor/mod.rs:257:19 [INFO] [stderr] | [INFO] [stderr] 257 | fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { [INFO] [stderr] | ^ help: if this is intentional, prefix it with an underscore: `_f` [INFO] [stderr] [INFO] [stderr] warning: unused variable: `start_idx` [INFO] [stderr] --> src/tensor/mod.rs:686:30 [INFO] [stderr] | [INFO] [stderr] 686 | fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec { [INFO] [stderr] | ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_start_idx` [INFO] [stderr] [INFO] [stderr] warning: unused variable: `size` [INFO] [stderr] --> src/tensor/mod.rs:686:48 [INFO] [stderr] | [INFO] [stderr] 686 | fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec { [INFO] [stderr] | ^^^^ help: if this is intentional, prefix it with an underscore: `_size` [INFO] [stderr] [INFO] [stderr] warning: unused variable: `rowcol` [INFO] [stderr] --> src/tensor/mod.rs:889:19 [INFO] [stderr] | [INFO] [stderr] 889 | fn sum(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stderr] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol` [INFO] [stderr] [INFO] [stderr] warning: unused variable: `dimension` [INFO] [stderr] --> src/tensor/mod.rs:889:34 [INFO] [stderr] | [INFO] [stderr] 889 | fn sum(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stderr] | ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension` [INFO] [stderr] [INFO] [stderr] warning: unused variable: `rowcol` [INFO] [stderr] --> src/tensor/mod.rs:904:20 [INFO] [stderr] | [INFO] [stderr] 904 | fn prod(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stderr] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol` [INFO] [stderr] [INFO] [stderr] warning: unused variable: `dimension` [INFO] [stderr] --> src/tensor/mod.rs:904:35 [INFO] [stderr] | [INFO] [stderr] 904 | fn prod(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stderr] | ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension` [INFO] [stderr] [INFO] [stderr] warning: field `layers` is never read [INFO] [stderr] --> src/nn/mod.rs:39:5 [INFO] [stderr] | [INFO] [stderr] 37 | pub struct Net { [INFO] [stderr] | --- field in this struct [INFO] [stderr] 38 | /// Represents all layers in the network [INFO] [stderr] 39 | layers: Vec, [INFO] [stderr] | ^^^^^^ [INFO] [stderr] | [INFO] [stderr] = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default [INFO] [stderr] [INFO] [stderr] warning: fields `data`, `batch_size`, and `shuffle` are never read [INFO] [stderr] --> src/nn/dataloader.rs:10:5 [INFO] [stderr] | [INFO] [stderr] 8 | pub struct DataLoader { [INFO] [stderr] | ---------- fields in this struct [INFO] [stderr] 9 | /// Dataset [INFO] [stderr] 10 | data: D, [INFO] [stderr] | ^^^^ [INFO] [stderr] 11 | /// Batch size [INFO] [stderr] 12 | batch_size: usize, [INFO] [stderr] | ^^^^^^^^^^ [INFO] [stderr] 13 | /// Whether or not to shuffle data [INFO] [stderr] 14 | shuffle: bool, [INFO] [stderr] | ^^^^^^^ [INFO] [stderr] | [INFO] [stderr] = note: `DataLoader` has a derived impl for the trait `Clone`, but this is intentionally ignored during dead code analysis [INFO] [stderr] [INFO] [stderr] warning: associated items `new` and `iter` are never used [INFO] [stderr] --> src/nn/dataloader.rs:41:8 [INFO] [stderr] | [INFO] [stderr] 33 | impl DataLoader { [INFO] [stderr] | -------------------------------------- associated items in this implementation [INFO] [stderr] ... [INFO] [stderr] 41 | fn new(data: D, batch_size: usize, shuffle: bool) -> DataLoader { [INFO] [stderr] | ^^^ [INFO] [stderr] ... [INFO] [stderr] 56 | fn iter(&mut self) -> DataLoader { [INFO] [stderr] | ^^^^ [INFO] [stderr] [INFO] [stderr] warning: fields `img_dir`, `transform`, and `target_transform` are never read [INFO] [stderr] --> src/nn/dataset.rs:33:5 [INFO] [stderr] | [INFO] [stderr] 31 | pub struct CustomImageDataset { [INFO] [stderr] | ------------------ fields in this struct [INFO] [stderr] 32 | img_labels: Vec, [INFO] [stderr] 33 | img_dir: &'static str, [INFO] [stderr] | ^^^^^^^ [INFO] [stderr] 34 | transform: Option, [INFO] [stderr] | ^^^^^^^^^ [INFO] [stderr] 35 | target_transform: Option, [INFO] [stderr] | ^^^^^^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: multiple fields are never read [INFO] [stderr] --> src/nn/optimizer.rs:38:5 [INFO] [stderr] | [INFO] [stderr] 36 | pub struct Adam { [INFO] [stderr] | ---- fields in this struct [INFO] [stderr] 37 | /// Learning rate [INFO] [stderr] 38 | lr: T, [INFO] [stderr] | ^^ [INFO] [stderr] 39 | /// Momentum [INFO] [stderr] 40 | momentum: f32, [INFO] [stderr] | ^^^^^^^^ [INFO] [stderr] 41 | /// Decay rate [INFO] [stderr] 42 | decay_rate: f32, [INFO] [stderr] | ^^^^^^^^^^ [INFO] [stderr] 43 | /// Beta 1 [INFO] [stderr] 44 | beta1: f32, [INFO] [stderr] | ^^^^^ [INFO] [stderr] 45 | /// Beta 2: [INFO] [stderr] 46 | beta2: f32, [INFO] [stderr] | ^^^^^ [INFO] [stderr] 47 | /// Epsilon [INFO] [stderr] 48 | epsilon: f32, [INFO] [stderr] | ^^^^^^^ [INFO] [stderr] 49 | m_dw: f32, [INFO] [stderr] | ^^^^ [INFO] [stderr] 50 | v_dw: f32, [INFO] [stderr] | ^^^^ [INFO] [stderr] 51 | m_db: f32, [INFO] [stderr] | ^^^^ [INFO] [stderr] 52 | v_db: f32, [INFO] [stderr] | ^^^^ [INFO] [stderr] [INFO] [stderr] warning: fields `lr`, `momentum`, and `decay_rate` are never read [INFO] [stderr] --> src/nn/optimizer.rs:97:5 [INFO] [stderr] | [INFO] [stderr] 95 | pub struct SGD { [INFO] [stderr] | --- fields in this struct [INFO] [stderr] 96 | /// Learning rate [INFO] [stderr] 97 | lr: T, [INFO] [stderr] | ^^ [INFO] [stderr] 98 | /// Momentum [INFO] [stderr] 99 | momentum: f32, [INFO] [stderr] | ^^^^^^^^ [INFO] [stderr] 100 | /// Decay rate [INFO] [stderr] 101 | decay_rate: f32, [INFO] [stderr] | ^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: function `swap` is never used [INFO] [stderr] --> src/tensor/mod.rs:33:4 [INFO] [stderr] | [INFO] [stderr] 33 | fn swap(lhs: &mut usize, rhs: &mut usize) { [INFO] [stderr] | ^^^^ [INFO] [stderr] [INFO] [stderr] warning: methods `get_sub_tensor`, `sum`, and `prod` are never used [INFO] [stderr] --> src/tensor/mod.rs:686:8 [INFO] [stderr] | [INFO] [stderr] 560 | / impl<'a, T> Tensor<'a, T> [INFO] [stderr] 561 | | where [INFO] [stderr] 562 | | T: TensorElement + Div + Sum, [INFO] [stderr] 563 | | ::Err: Error + 'static, [INFO] [stderr] 564 | | Vec: IntoParallelIterator, [INFO] [stderr] 565 | | Vec<&'a T>: IntoParallelRefIterator<'a>, [INFO] [stderr] | |____________________________________________- methods in this implementation [INFO] [stderr] ... [INFO] [stderr] 686 | fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec { [INFO] [stderr] | ^^^^^^^^^^^^^^ [INFO] [stderr] ... [INFO] [stderr] 889 | fn sum(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stderr] | ^^^ [INFO] [stderr] ... [INFO] [stderr] 904 | fn prod(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stderr] | ^^^^ [INFO] [stderr] [INFO] [stderr] warning: function `ReLU` should have a snake case name [INFO] [stderr] --> src/nn/activation.rs:25:8 [INFO] [stderr] | [INFO] [stderr] 25 | pub fn ReLU<'a, 'b, T>(x: &Tensor<'a, T>) -> Tensor<'b, T> [INFO] [stderr] | ^^^^ help: convert the identifier to snake case: `re_lu` [INFO] [stderr] | [INFO] [stderr] = note: `#[warn(non_snake_case)]` (part of `#[warn(nonstandard_style)]`) on by default [INFO] [stderr] [INFO] [stderr] warning: function `PReLU` should have a snake case name [INFO] [stderr] --> src/nn/activation.rs:52:8 [INFO] [stderr] | [INFO] [stderr] 52 | pub fn PReLU<'a, 'b, T>(x: &Tensor<'a, T>, alpha: T) -> Tensor<'b, T> [INFO] [stderr] | ^^^^^ help: convert the identifier to snake case: `pre_lu` [INFO] [stderr] [INFO] [stderr] warning: function `Sigmoid` should have a snake case name [INFO] [stderr] --> src/nn/activation.rs:79:8 [INFO] [stderr] | [INFO] [stderr] 79 | pub fn Sigmoid<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T> [INFO] [stderr] | ^^^^^^^ help: convert the identifier to snake case (notice the capitalization): `sigmoid` [INFO] [stderr] [INFO] [stderr] warning: function `GeLU` should have a snake case name [INFO] [stderr] --> src/nn/activation.rs:108:8 [INFO] [stderr] | [INFO] [stderr] 108 | pub fn GeLU<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T> [INFO] [stderr] | ^^^^ help: convert the identifier to snake case: `ge_lu` [INFO] [stderr] [INFO] [stderr] warning: missing documentation for an associated function [INFO] [stderr] --> src/nn/loss.rs:14:5 [INFO] [stderr] | [INFO] [stderr] 14 | fn loss(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T; [INFO] [stderr] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stderr] | [INFO] [stderr] note: the lint level is defined here [INFO] [stderr] --> src/nn/loss.rs:2:9 [INFO] [stderr] | [INFO] [stderr] 2 | #![warn(missing_docs)] [INFO] [stderr] | ^^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: missing documentation for an associated function [INFO] [stderr] --> src/nn/loss.rs:15:5 [INFO] [stderr] | [INFO] [stderr] 15 | fn backward(); [INFO] [stderr] | ^^^^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: function `BCEntroypyLoss` should have a snake case name [INFO] [stderr] --> src/nn/loss.rs:31:8 [INFO] [stderr] | [INFO] [stderr] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stderr] | ^^^^^^^^^^^^^^ help: convert the identifier to snake case: `bcentroypy_loss` [INFO] [stderr] [INFO] [stderr] warning: function `CEntroypyLoss` should have a snake case name [INFO] [stderr] --> src/nn/loss.rs:50:8 [INFO] [stderr] | [INFO] [stderr] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stderr] | ^^^^^^^^^^^^^ help: convert the identifier to snake case: `centroypy_loss` [INFO] [stderr] [INFO] [stderr] warning: function `L1Loss` should have a snake case name [INFO] [stderr] --> src/nn/loss.rs:73:8 [INFO] [stderr] | [INFO] [stderr] 73 | pub fn L1Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stderr] | ^^^^^^ help: convert the identifier to snake case: `l1_loss` [INFO] [stderr] [INFO] [stderr] warning: function `L2Loss` should have a snake case name [INFO] [stderr] --> src/nn/loss.rs:96:8 [INFO] [stderr] | [INFO] [stderr] 96 | pub fn L2Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stderr] | ^^^^^^ help: convert the identifier to snake case: `l2_loss` [INFO] [stderr] [INFO] [stderr] warning: function `MaxPool` should have a snake case name [INFO] [stderr] --> src/nn/pooling.rs:23:8 [INFO] [stderr] | [INFO] [stderr] 23 | pub fn MaxPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T> [INFO] [stderr] | ^^^^^^^ help: convert the identifier to snake case: `max_pool` [INFO] [stderr] [INFO] [stderr] warning: function `MinPool` should have a snake case name [INFO] [stderr] --> src/nn/pooling.rs:62:8 [INFO] [stderr] | [INFO] [stderr] 62 | pub fn MinPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T> [INFO] [stderr] | ^^^^^^^ help: convert the identifier to snake case: `min_pool` [INFO] [stderr] [INFO] [stderr] warning: function `AvgPool` should have a snake case name [INFO] [stderr] --> src/nn/pooling.rs:93:8 [INFO] [stderr] | [INFO] [stderr] 93 | pub fn AvgPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T> [INFO] [stderr] | ^^^^^^^ help: convert the identifier to snake case: `avg_pool` [INFO] [stderr] [INFO] [stderr] warning: missing documentation for a struct [INFO] [stderr] --> src/nn/transform.rs:4:1 [INFO] [stderr] | [INFO] [stderr] 4 | pub struct Transform; [INFO] [stderr] | ^^^^^^^^^^^^^^^^^^^^ [INFO] [stderr] | [INFO] [stderr] note: the lint level is defined here [INFO] [stderr] --> src/nn/transform.rs:2:9 [INFO] [stderr] | [INFO] [stderr] 2 | #![warn(missing_docs)] [INFO] [stderr] | ^^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: missing documentation for a struct [INFO] [stderr] --> src/tensor/mod.rs:145:1 [INFO] [stderr] | [INFO] [stderr] 145 | pub struct Tensor<'a, T> [INFO] [stderr] | ^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stderr] | [INFO] [stderr] note: the lint level is defined here [INFO] [stderr] --> src/lib.rs:1:9 [INFO] [stderr] | [INFO] [stderr] 1 | #![warn(missing_docs)] [INFO] [stderr] | ^^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: `kaffe` (lib) generated 45 warnings (run `cargo fix --lib -p kaffe` to apply 22 suggestions) [INFO] [stderr] warning: `kaffe` (lib test) generated 45 warnings (45 duplicates) [INFO] [stderr] Finished `test` profile [unoptimized + debuginfo] target(s) in 0.11s [INFO] [stderr] Running unittests src/lib.rs (/opt/rustwide/target/debug/deps/kaffe-97bede8cfcb8b023) [INFO] [stdout] [INFO] [stdout] running 0 tests [INFO] [stdout] [INFO] [stdout] test result: ok. 0 passed; 0 failed; 0 ignored; 0 measured; 0 filtered out; finished in 0.00s [INFO] [stdout] [INFO] [stderr] Running tests/nn.rs (/opt/rustwide/target/debug/deps/nn-1c88b5395785526c) [INFO] [stdout] [INFO] [stdout] running 1 test [INFO] [stdout] test relu ... ok [INFO] [stdout] [INFO] [stdout] test result: ok. 1 passed; 0 failed; 0 ignored; 0 measured; 0 filtered out; finished in 0.00s [INFO] [stdout] [INFO] [stderr] Running tests/tensor.rs (/opt/rustwide/target/debug/deps/tensor-ccf94bfb46ee0ec6) [INFO] [stdout] [INFO] [stdout] running 2 tests [INFO] [stderr] Doc-tests kaffe [INFO] [stdout] test creation ... ok [INFO] [stdout] test matmul ... ok [INFO] [stdout] [INFO] [stdout] test result: ok. 2 passed; 0 failed; 0 ignored; 0 measured; 0 filtered out; finished in 0.01s [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] running 85 tests [INFO] [stdout] test src/nn/dataloader.rs - nn::dataloader::DataLoader::new (line 38) ... ok [INFO] [stdout] test src/nn/dataloader.rs - nn::dataloader::DataLoader::next (line 25) ... ok [INFO] [stdout] test src/nn/optimizer.rs - nn::optimizer::Adam (line 28) ... ok [INFO] [stdout] test src/nn/activation.rs - nn::activation::GeLU (line 100) ... ok [INFO] [stdout] test src/nn/dataloader.rs - nn::dataloader::DataLoader::iter (line 53) ... ok [INFO] [stdout] test src/nn/loss.rs - nn::loss::BCEntroypyLoss (line 24) ... ok [INFO] [stdout] test src/nn/optimizer.rs - nn::optimizer::SGD (line 87) ... ok [INFO] [stdout] test src/nn/activation.rs - nn::activation::Sigmoid (line 71) ... ok [INFO] [stdout] test src/nn/loss.rs - nn::loss::L1Loss (line 64) ... ok [INFO] [stdout] test src/nn/pooling.rs - nn::pooling::MaxPool (line 12) ... ok [INFO] [stdout] test src/nn/loss.rs - nn::loss::CEntroypyLoss (line 43) ... ok [INFO] [stdout] test src/nn/activation.rs - nn::activation::ReLU (line 17) ... ok [INFO] [stdout] test src/nn/pooling.rs - nn::pooling::AvgPool (line 81) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::add (line 922) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::cosh (line 1209) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::all (line 1693) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::abs_self (line 1348) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::any (line 1675) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::abs (line 1245) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::avg (line 808) ... ok [INFO] [stdout] test src/nn/activation.rs - nn::activation::PReLU (line 44) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::cumprod (line 793) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::add_self (line 1264) ... ok [INFO] [stdout] test src/nn/pooling.rs - nn::pooling::MinPool (line 42) ... ok [INFO] [stdout] test src/nn/loss.rs - nn::loss::L2Loss (line 87) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::count_where (line 1606) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::add_val (line 1061) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::cumsum (line 778) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::default (line 273) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::add_val_self (line 1365) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::div_val_self (line 1419) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::get_sub_tensor (line 681) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::find (line 1711) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::from_file (line 526) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::eye (line 341) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::get_vec_slice (line 656) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::identity (line 362) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::div_self (line 1327) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::from_slice (line 378) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::init (line 324) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::eigenvalue (line 1582) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::log (line 1136) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::get (line 628) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::ln (line 1156) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::div (line 1030) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::find_all (line 1733) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::max (line 737) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::mul (line 1003) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::median (line 844) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::div_val (line 1112) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::matmul (line 1437) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::new (line 298) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::min (line 757) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::mean (line 829) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::ones (line 417) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::mul_self (line 1306) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::prod (line 897) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::mul_val_self (line 1401) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::random_like (line 470) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::randomize_range (line 487) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::mul_val (line 1095) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::randomize (line 511) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::ones_like (line 452) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::set (line 695) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::sinh (line 1192) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::set_many (line 719) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::pow (line 1226) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::sub_abs (line 976) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::set_where (line 1646) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::squeeze (line 593) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::mm (line 1472) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::zeros (line 400) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::sum (line 883) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::reshape (line 572) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::size (line 611) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::transpose_copy (line 1564) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::t (line 1548) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::sub_val_self (line 1383) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::zeros_like (line 434) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::transpose (line 1513) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::tanh (line 1175) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::sum_where (line 1624) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::sub_self (line 1285) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::sub (line 949) ... ok [INFO] [stdout] test src/tensor/mod.rs - tensor::Tensor<'a,T>::sub_val (line 1078) ... ok [INFO] [stdout] [INFO] [stdout] test result: ok. 85 passed; 0 failed; 0 ignored; 0 measured; 0 filtered out; finished in 7.93s [INFO] [stdout] [INFO] running `Command { std: "docker" "inspect" "4b0eda8e16ea00dc4c4f01466c4c93aae3c72defb22f1eace42fe9bbe2ff9fd5", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "4b0eda8e16ea00dc4c4f01466c4c93aae3c72defb22f1eace42fe9bbe2ff9fd5", kill_on_drop: false }` [INFO] [stdout] 4b0eda8e16ea00dc4c4f01466c4c93aae3c72defb22f1eace42fe9bbe2ff9fd5