[INFO] fetching crate kaffe 0.2.0... [INFO] checking kaffe-0.2.0 against try#a4980311fb7bb9e7893708e6bd3fbbfb2819fd3d for pr-145342-1 [INFO] extracting crate kaffe 0.2.0 into /workspace/builds/worker-2-tc2/source [INFO] started tweaking crates.io crate kaffe 0.2.0 [INFO] finished tweaking crates.io crate kaffe 0.2.0 [INFO] tweaked toml for crates.io crate kaffe 0.2.0 written to /workspace/builds/worker-2-tc2/source/Cargo.toml [INFO] validating manifest of crates.io crate kaffe 0.2.0 on toolchain a4980311fb7bb9e7893708e6bd3fbbfb2819fd3d [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+a4980311fb7bb9e7893708e6bd3fbbfb2819fd3d" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }` [INFO] crate crates.io crate kaffe 0.2.0 already has a lockfile, it will not be regenerated [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+a4980311fb7bb9e7893708e6bd3fbbfb2819fd3d" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }` [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-2-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-2-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:7ad1b28ee6f5f7f699f6cf7015098d6ccdd96d6f2d78dd06228f5b4c9faf309c" "/opt/rustwide/cargo-home/bin/cargo" "+a4980311fb7bb9e7893708e6bd3fbbfb2819fd3d" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }` [INFO] [stdout] f1c708593eca087c837bdb7312e36acae140bb7daeef4d4116e725ec08ef5cce [INFO] running `Command { std: "docker" "start" "-a" "f1c708593eca087c837bdb7312e36acae140bb7daeef4d4116e725ec08ef5cce", kill_on_drop: false }` [INFO] running `Command { std: "docker" "inspect" "f1c708593eca087c837bdb7312e36acae140bb7daeef4d4116e725ec08ef5cce", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "f1c708593eca087c837bdb7312e36acae140bb7daeef4d4116e725ec08ef5cce", kill_on_drop: false }` [INFO] [stdout] f1c708593eca087c837bdb7312e36acae140bb7daeef4d4116e725ec08ef5cce [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-2-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-2-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:7ad1b28ee6f5f7f699f6cf7015098d6ccdd96d6f2d78dd06228f5b4c9faf309c" "/opt/rustwide/cargo-home/bin/cargo" "+a4980311fb7bb9e7893708e6bd3fbbfb2819fd3d" "check" "--frozen" "--all" "--all-targets" "--message-format=json", kill_on_drop: false }` [INFO] [stdout] 61f0b901987c926f35d5ed5b191c927f8576776347465c5cf242ad49967a46d5 [INFO] running `Command { std: "docker" "start" "-a" "61f0b901987c926f35d5ed5b191c927f8576776347465c5cf242ad49967a46d5", kill_on_drop: false }` [INFO] [stderr] Compiling autocfg v1.1.0 [INFO] [stderr] Compiling libc v0.2.146 [INFO] [stderr] Checking cfg-if v1.0.0 [INFO] [stderr] Compiling proc-macro2 v1.0.60 [INFO] [stderr] Compiling quote v1.0.28 [INFO] [stderr] Compiling unicode-ident v1.0.9 [INFO] [stderr] Compiling crossbeam-utils v0.8.15 [INFO] [stderr] Compiling serde v1.0.164 [INFO] [stderr] Checking scopeguard v1.1.0 [INFO] [stderr] Compiling rayon-core v1.11.0 [INFO] [stderr] Checking either v1.8.1 [INFO] [stderr] Compiling io-lifetimes v1.0.11 [INFO] [stderr] Checking bitflags v1.3.2 [INFO] [stderr] Compiling serde_json v1.0.96 [INFO] [stderr] Compiling rustix v0.37.19 [INFO] [stderr] Checking itertools v0.10.5 [INFO] [stderr] Checking clap_lex v0.5.0 [INFO] [stderr] Compiling memoffset v0.8.0 [INFO] [stderr] Compiling crossbeam-epoch v0.9.14 [INFO] [stderr] Compiling num-traits v0.2.15 [INFO] [stderr] Checking anstyle v1.0.0 [INFO] [stderr] Checking crossbeam-channel v0.5.8 [INFO] [stderr] Checking ryu v1.0.13 [INFO] [stderr] Compiling anyhow v1.0.71 [INFO] [stderr] Checking itoa v1.0.6 [INFO] [stderr] Checking ciborium-io v0.2.1 [INFO] [stderr] Checking crossbeam-deque v0.8.3 [INFO] [stderr] Checking plotters-backend v0.3.4 [INFO] [stderr] Checking linux-raw-sys v0.3.8 [INFO] [stderr] Checking half v1.8.2 [INFO] [stderr] Checking ppv-lite86 v0.2.17 [INFO] [stderr] Checking ciborium-ll v0.2.1 [INFO] [stderr] Compiling syn v2.0.18 [INFO] [stderr] Checking plotters-svg v0.3.3 [INFO] [stderr] Checking clap_builder v4.3.3 [INFO] [stderr] Checking cast v0.3.0 [INFO] [stderr] Checking same-file v1.0.6 [INFO] [stderr] Checking regex-syntax v0.7.2 [INFO] [stderr] Checking walkdir v2.3.3 [INFO] [stderr] Checking once_cell v1.18.0 [INFO] [stderr] Checking criterion-plot v0.5.0 [INFO] [stderr] Checking plotters v0.3.4 [INFO] [stderr] Checking oorandom v11.1.3 [INFO] [stderr] Checking anes v0.1.6 [INFO] [stderr] Checking num_cpus v1.15.0 [INFO] [stderr] Checking getrandom v0.2.10 [INFO] [stderr] Checking rand_core v0.6.4 [INFO] [stderr] Checking rand_chacha v0.3.1 [INFO] [stderr] Checking rand v0.8.5 [INFO] [stderr] Checking rayon v1.7.0 [INFO] [stderr] Checking regex v1.8.4 [INFO] [stderr] Checking clap v4.3.3 [INFO] [stderr] Checking is-terminal v0.4.7 [INFO] [stderr] Compiling serde_derive v1.0.164 [INFO] [stderr] Checking ciborium v0.2.1 [INFO] [stderr] Checking kaffe v0.2.0 (/opt/rustwide/workdir) [INFO] [stdout] warning: unused import: `marker::PhantomData` [INFO] [stdout] --> src/nn/optimizer.rs:4:25 [INFO] [stdout] | [INFO] [stdout] 4 | use std::{error::Error, marker::PhantomData, str::FromStr}; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `rayon::prelude::*` [INFO] [stdout] --> src/nn/optimizer.rs:7:5 [INFO] [stdout] | [INFO] [stdout] 7 | use rayon::prelude::*; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `idx` [INFO] [stdout] --> src/nn/dataset.rs:57:26 [INFO] [stdout] | [INFO] [stdout] 57 | fn get<'a, T>(&self, idx: usize) -> (Tensor<'a, T>, String) [INFO] [stdout] | ^^^ help: if this is intentional, prefix it with an underscore: `_idx` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_variables)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `y` [INFO] [stdout] --> src/nn/loss.rs:31:30 [INFO] [stdout] | [INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_y` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `y_hat` [INFO] [stdout] --> src/nn/loss.rs:31:49 [INFO] [stdout] | [INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `y` [INFO] [stdout] --> src/nn/loss.rs:50:29 [INFO] [stdout] | [INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_y` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `y_hat` [INFO] [stdout] --> src/nn/loss.rs:50:48 [INFO] [stdout] | [INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `cost` [INFO] [stdout] --> src/nn/optimizer.rs:75:31 [INFO] [stdout] | [INFO] [stdout] 75 | fn minimize(&mut self, cost: F, vars: &mut Vec) [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_cost` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `vars` [INFO] [stdout] --> src/nn/optimizer.rs:75:40 [INFO] [stdout] | [INFO] [stdout] 75 | fn minimize(&mut self, cost: F, vars: &mut Vec) [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_vars` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `cost` [INFO] [stdout] --> src/nn/optimizer.rs:117:31 [INFO] [stdout] | [INFO] [stdout] 117 | fn minimize(&mut self, cost: F, vars: &mut Vec) [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_cost` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `vars` [INFO] [stdout] --> src/nn/optimizer.rs:117:40 [INFO] [stdout] | [INFO] [stdout] 117 | fn minimize(&mut self, cost: F, vars: &mut Vec) [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_vars` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `tensor` [INFO] [stdout] --> src/nn/mod.rs:24:5 [INFO] [stdout] | [INFO] [stdout] 24 | tensor: Tensor<'a, T>, [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `kernel` [INFO] [stdout] --> src/nn/mod.rs:25:5 [INFO] [stdout] | [INFO] [stdout] 25 | kernel: Tensor<'a, T>, [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_kernel` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `stride` [INFO] [stdout] --> src/nn/mod.rs:26:5 [INFO] [stdout] | [INFO] [stdout] 26 | stride: usize, [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_stride` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `padding` [INFO] [stdout] --> src/nn/mod.rs:27:5 [INFO] [stdout] | [INFO] [stdout] 27 | padding: usize, [INFO] [stdout] | ^^^^^^^ help: if this is intentional, prefix it with an underscore: `_padding` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `f` [INFO] [stdout] --> src/tensor/mod.rs:257:19 [INFO] [stdout] | [INFO] [stdout] 257 | fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_f` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `start_idx` [INFO] [stdout] --> src/tensor/mod.rs:686:30 [INFO] [stdout] | [INFO] [stdout] 686 | fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec { [INFO] [stdout] | ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_start_idx` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `size` [INFO] [stdout] --> src/tensor/mod.rs:686:48 [INFO] [stdout] | [INFO] [stdout] 686 | fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec { [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_size` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `rowcol` [INFO] [stdout] --> src/tensor/mod.rs:889:19 [INFO] [stdout] | [INFO] [stdout] 889 | fn sum(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `dimension` [INFO] [stdout] --> src/tensor/mod.rs:889:34 [INFO] [stdout] | [INFO] [stdout] 889 | fn sum(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `rowcol` [INFO] [stdout] --> src/tensor/mod.rs:904:20 [INFO] [stdout] | [INFO] [stdout] 904 | fn prod(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `dimension` [INFO] [stdout] --> src/tensor/mod.rs:904:35 [INFO] [stdout] | [INFO] [stdout] 904 | fn prod(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `layers` is never read [INFO] [stdout] --> src/nn/mod.rs:39:5 [INFO] [stdout] | [INFO] [stdout] 37 | pub struct Net { [INFO] [stdout] | --- field in this struct [INFO] [stdout] 38 | /// Represents all layers in the network [INFO] [stdout] 39 | layers: Vec, [INFO] [stdout] | ^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(dead_code)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: fields `data`, `batch_size`, and `shuffle` are never read [INFO] [stdout] --> src/nn/dataloader.rs:10:5 [INFO] [stdout] | [INFO] [stdout] 8 | pub struct DataLoader { [INFO] [stdout] | ---------- fields in this struct [INFO] [stdout] 9 | /// Dataset [INFO] [stdout] 10 | data: D, [INFO] [stdout] | ^^^^ [INFO] [stdout] 11 | /// Batch size [INFO] [stdout] 12 | batch_size: usize, [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] 13 | /// Whether or not to shuffle data [INFO] [stdout] 14 | shuffle: bool, [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `DataLoader` has a derived impl for the trait `Clone`, but this is intentionally ignored during dead code analysis [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: associated items `new` and `iter` are never used [INFO] [stdout] --> src/nn/dataloader.rs:41:8 [INFO] [stdout] | [INFO] [stdout] 33 | impl DataLoader { [INFO] [stdout] | -------------------------------------- associated items in this implementation [INFO] [stdout] ... [INFO] [stdout] 41 | fn new(data: D, batch_size: usize, shuffle: bool) -> DataLoader { [INFO] [stdout] | ^^^ [INFO] [stdout] ... [INFO] [stdout] 56 | fn iter(&mut self) -> DataLoader { [INFO] [stdout] | ^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: fields `img_dir`, `transform`, and `target_transform` are never read [INFO] [stdout] --> src/nn/dataset.rs:33:5 [INFO] [stdout] | [INFO] [stdout] 31 | pub struct CustomImageDataset { [INFO] [stdout] | ------------------ fields in this struct [INFO] [stdout] 32 | img_labels: Vec, [INFO] [stdout] 33 | img_dir: &'static str, [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] 34 | transform: Option, [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] 35 | target_transform: Option, [INFO] [stdout] | ^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: multiple fields are never read [INFO] [stdout] --> src/nn/optimizer.rs:38:5 [INFO] [stdout] | [INFO] [stdout] 36 | pub struct Adam { [INFO] [stdout] | ---- fields in this struct [INFO] [stdout] 37 | /// Learning rate [INFO] [stdout] 38 | lr: T, [INFO] [stdout] | ^^ [INFO] [stdout] 39 | /// Momentum [INFO] [stdout] 40 | momentum: f32, [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] 41 | /// Decay rate [INFO] [stdout] 42 | decay_rate: f32, [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] 43 | /// Beta 1 [INFO] [stdout] 44 | beta1: f32, [INFO] [stdout] | ^^^^^ [INFO] [stdout] 45 | /// Beta 2: [INFO] [stdout] 46 | beta2: f32, [INFO] [stdout] | ^^^^^ [INFO] [stdout] 47 | /// Epsilon [INFO] [stdout] 48 | epsilon: f32, [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] 49 | m_dw: f32, [INFO] [stdout] | ^^^^ [INFO] [stdout] 50 | v_dw: f32, [INFO] [stdout] | ^^^^ [INFO] [stdout] 51 | m_db: f32, [INFO] [stdout] | ^^^^ [INFO] [stdout] 52 | v_db: f32, [INFO] [stdout] | ^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: fields `lr`, `momentum`, and `decay_rate` are never read [INFO] [stdout] --> src/nn/optimizer.rs:97:5 [INFO] [stdout] | [INFO] [stdout] 95 | pub struct SGD { [INFO] [stdout] | --- fields in this struct [INFO] [stdout] 96 | /// Learning rate [INFO] [stdout] 97 | lr: T, [INFO] [stdout] | ^^ [INFO] [stdout] 98 | /// Momentum [INFO] [stdout] 99 | momentum: f32, [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] 100 | /// Decay rate [INFO] [stdout] 101 | decay_rate: f32, [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `swap` is never used [INFO] [stdout] --> src/tensor/mod.rs:33:4 [INFO] [stdout] | [INFO] [stdout] 33 | fn swap(lhs: &mut usize, rhs: &mut usize) { [INFO] [stdout] | ^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: methods `get_sub_tensor`, `sum`, and `prod` are never used [INFO] [stdout] --> src/tensor/mod.rs:686:8 [INFO] [stdout] | [INFO] [stdout] 560 | / impl<'a, T> Tensor<'a, T> [INFO] [stdout] 561 | | where [INFO] [stdout] 562 | | T: TensorElement + Div + Sum, [INFO] [stdout] 563 | | ::Err: Error + 'static, [INFO] [stdout] 564 | | Vec: IntoParallelIterator, [INFO] [stdout] 565 | | Vec<&'a T>: IntoParallelRefIterator<'a>, [INFO] [stdout] | |____________________________________________- methods in this implementation [INFO] [stdout] ... [INFO] [stdout] 686 | fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec { [INFO] [stdout] | ^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 889 | fn sum(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^ [INFO] [stdout] ... [INFO] [stdout] 904 | fn prod(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `ReLU` should have a snake case name [INFO] [stdout] --> src/nn/activation.rs:25:8 [INFO] [stdout] | [INFO] [stdout] 25 | pub fn ReLU<'a, 'b, T>(x: &Tensor<'a, T>) -> Tensor<'b, T> [INFO] [stdout] | ^^^^ help: convert the identifier to snake case: `re_lu` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(non_snake_case)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `PReLU` should have a snake case name [INFO] [stdout] --> src/nn/activation.rs:52:8 [INFO] [stdout] | [INFO] [stdout] 52 | pub fn PReLU<'a, 'b, T>(x: &Tensor<'a, T>, alpha: T) -> Tensor<'b, T> [INFO] [stdout] | ^^^^^ help: convert the identifier to snake case: `pre_lu` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `Sigmoid` should have a snake case name [INFO] [stdout] --> src/nn/activation.rs:79:8 [INFO] [stdout] | [INFO] [stdout] 79 | pub fn Sigmoid<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T> [INFO] [stdout] | ^^^^^^^ help: convert the identifier to snake case (notice the capitalization): `sigmoid` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `GeLU` should have a snake case name [INFO] [stdout] --> src/nn/activation.rs:108:8 [INFO] [stdout] | [INFO] [stdout] 108 | pub fn GeLU<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T> [INFO] [stdout] | ^^^^ help: convert the identifier to snake case: `ge_lu` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: missing documentation for an associated function [INFO] [stdout] --> src/nn/loss.rs:14:5 [INFO] [stdout] | [INFO] [stdout] 14 | fn loss(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] note: the lint level is defined here [INFO] [stdout] --> src/nn/loss.rs:2:9 [INFO] [stdout] | [INFO] [stdout] 2 | #![warn(missing_docs)] [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: missing documentation for an associated function [INFO] [stdout] --> src/nn/loss.rs:15:5 [INFO] [stdout] | [INFO] [stdout] 15 | fn backward(); [INFO] [stdout] | ^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `BCEntroypyLoss` should have a snake case name [INFO] [stdout] --> src/nn/loss.rs:31:8 [INFO] [stdout] | [INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^^^^^^^^^^ help: convert the identifier to snake case: `bcentroypy_loss` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `CEntroypyLoss` should have a snake case name [INFO] [stdout] --> src/nn/loss.rs:50:8 [INFO] [stdout] | [INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^^^^^^^^^ help: convert the identifier to snake case: `centroypy_loss` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `L1Loss` should have a snake case name [INFO] [stdout] --> src/nn/loss.rs:73:8 [INFO] [stdout] | [INFO] [stdout] 73 | pub fn L1Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^^ help: convert the identifier to snake case: `l1_loss` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `L2Loss` should have a snake case name [INFO] [stdout] --> src/nn/loss.rs:96:8 [INFO] [stdout] | [INFO] [stdout] 96 | pub fn L2Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^^ help: convert the identifier to snake case: `l2_loss` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `MaxPool` should have a snake case name [INFO] [stdout] --> src/nn/pooling.rs:23:8 [INFO] [stdout] | [INFO] [stdout] 23 | pub fn MaxPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T> [INFO] [stdout] | ^^^^^^^ help: convert the identifier to snake case: `max_pool` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `MinPool` should have a snake case name [INFO] [stdout] --> src/nn/pooling.rs:62:8 [INFO] [stdout] | [INFO] [stdout] 62 | pub fn MinPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T> [INFO] [stdout] | ^^^^^^^ help: convert the identifier to snake case: `min_pool` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `AvgPool` should have a snake case name [INFO] [stdout] --> src/nn/pooling.rs:93:8 [INFO] [stdout] | [INFO] [stdout] 93 | pub fn AvgPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T> [INFO] [stdout] | ^^^^^^^ help: convert the identifier to snake case: `avg_pool` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: missing documentation for a struct [INFO] [stdout] --> src/nn/transform.rs:4:1 [INFO] [stdout] | [INFO] [stdout] 4 | pub struct Transform; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] note: the lint level is defined here [INFO] [stdout] --> src/nn/transform.rs:2:9 [INFO] [stdout] | [INFO] [stdout] 2 | #![warn(missing_docs)] [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: missing documentation for a struct [INFO] [stdout] --> src/tensor/mod.rs:145:1 [INFO] [stdout] | [INFO] [stdout] 145 | pub struct Tensor<'a, T> [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] note: the lint level is defined here [INFO] [stdout] --> src/lib.rs:1:9 [INFO] [stdout] | [INFO] [stdout] 1 | #![warn(missing_docs)] [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] Checking tinytemplate v1.2.1 [INFO] [stderr] Checking criterion v0.5.1 [INFO] [stdout] error[E0432]: unresolved imports `kaffe::Matrix`, `kaffe::MatrixLinAlg` [INFO] [stdout] --> benches/matmul_benchmark.rs:1:13 [INFO] [stdout] | [INFO] [stdout] 1 | use kaffe::{Matrix, MatrixLinAlg}; [INFO] [stdout] | ^^^^^^ ^^^^^^^^^^^^ no `MatrixLinAlg` in the root [INFO] [stdout] | | [INFO] [stdout] | no `Matrix` in the root [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0432]: unresolved imports `kaffe::Matrix`, `kaffe::MatrixLinAlg` [INFO] [stdout] --> benches/matrix_ops_benches.rs:1:13 [INFO] [stdout] | [INFO] [stdout] 1 | use kaffe::{Matrix, MatrixLinAlg}; [INFO] [stdout] | ^^^^^^ ^^^^^^^^^^^^ no `MatrixLinAlg` in the root [INFO] [stdout] | | [INFO] [stdout] | no `Matrix` in the root [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] For more information about this error, try `rustc --explain E0432`. [INFO] [stdout] [INFO] [stdout] For more information about this error, try `rustc --explain E0432`. [INFO] [stdout] [INFO] [stderr] error: could not compile `kaffe` (bench "matmul_benchmark") due to 1 previous error [INFO] [stderr] warning: build failed, waiting for other jobs to finish... [INFO] [stderr] error: could not compile `kaffe` (bench "matrix_ops_benches") due to 1 previous error [INFO] [stdout] warning: unused import: `marker::PhantomData` [INFO] [stdout] --> src/nn/optimizer.rs:4:25 [INFO] [stdout] | [INFO] [stdout] 4 | use std::{error::Error, marker::PhantomData, str::FromStr}; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `rayon::prelude::*` [INFO] [stdout] --> src/nn/optimizer.rs:7:5 [INFO] [stdout] | [INFO] [stdout] 7 | use rayon::prelude::*; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `idx` [INFO] [stdout] --> src/nn/dataset.rs:57:26 [INFO] [stdout] | [INFO] [stdout] 57 | fn get<'a, T>(&self, idx: usize) -> (Tensor<'a, T>, String) [INFO] [stdout] | ^^^ help: if this is intentional, prefix it with an underscore: `_idx` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_variables)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `y` [INFO] [stdout] --> src/nn/loss.rs:31:30 [INFO] [stdout] | [INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_y` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `y_hat` [INFO] [stdout] --> src/nn/loss.rs:31:49 [INFO] [stdout] | [INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `y` [INFO] [stdout] --> src/nn/loss.rs:50:29 [INFO] [stdout] | [INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_y` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `y_hat` [INFO] [stdout] --> src/nn/loss.rs:50:48 [INFO] [stdout] | [INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_y_hat` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `cost` [INFO] [stdout] --> src/nn/optimizer.rs:75:31 [INFO] [stdout] | [INFO] [stdout] 75 | fn minimize(&mut self, cost: F, vars: &mut Vec) [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_cost` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `vars` [INFO] [stdout] --> src/nn/optimizer.rs:75:40 [INFO] [stdout] | [INFO] [stdout] 75 | fn minimize(&mut self, cost: F, vars: &mut Vec) [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_vars` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `cost` [INFO] [stdout] --> src/nn/optimizer.rs:117:31 [INFO] [stdout] | [INFO] [stdout] 117 | fn minimize(&mut self, cost: F, vars: &mut Vec) [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_cost` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `vars` [INFO] [stdout] --> src/nn/optimizer.rs:117:40 [INFO] [stdout] | [INFO] [stdout] 117 | fn minimize(&mut self, cost: F, vars: &mut Vec) [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_vars` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `tensor` [INFO] [stdout] --> src/nn/mod.rs:24:5 [INFO] [stdout] | [INFO] [stdout] 24 | tensor: Tensor<'a, T>, [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `kernel` [INFO] [stdout] --> src/nn/mod.rs:25:5 [INFO] [stdout] | [INFO] [stdout] 25 | kernel: Tensor<'a, T>, [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_kernel` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `stride` [INFO] [stdout] --> src/nn/mod.rs:26:5 [INFO] [stdout] | [INFO] [stdout] 26 | stride: usize, [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_stride` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `padding` [INFO] [stdout] --> src/nn/mod.rs:27:5 [INFO] [stdout] | [INFO] [stdout] 27 | padding: usize, [INFO] [stdout] | ^^^^^^^ help: if this is intentional, prefix it with an underscore: `_padding` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `f` [INFO] [stdout] --> src/tensor/mod.rs:257:19 [INFO] [stdout] | [INFO] [stdout] 257 | fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_f` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `start_idx` [INFO] [stdout] --> src/tensor/mod.rs:686:30 [INFO] [stdout] | [INFO] [stdout] 686 | fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec { [INFO] [stdout] | ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_start_idx` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `size` [INFO] [stdout] --> src/tensor/mod.rs:686:48 [INFO] [stdout] | [INFO] [stdout] 686 | fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec { [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_size` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `rowcol` [INFO] [stdout] --> src/tensor/mod.rs:889:19 [INFO] [stdout] | [INFO] [stdout] 889 | fn sum(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `dimension` [INFO] [stdout] --> src/tensor/mod.rs:889:34 [INFO] [stdout] | [INFO] [stdout] 889 | fn sum(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `rowcol` [INFO] [stdout] --> src/tensor/mod.rs:904:20 [INFO] [stdout] | [INFO] [stdout] 904 | fn prod(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_rowcol` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `dimension` [INFO] [stdout] --> src/tensor/mod.rs:904:35 [INFO] [stdout] | [INFO] [stdout] 904 | fn prod(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_dimension` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `layers` is never read [INFO] [stdout] --> src/nn/mod.rs:39:5 [INFO] [stdout] | [INFO] [stdout] 37 | pub struct Net { [INFO] [stdout] | --- field in this struct [INFO] [stdout] 38 | /// Represents all layers in the network [INFO] [stdout] 39 | layers: Vec, [INFO] [stdout] | ^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(dead_code)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: fields `data`, `batch_size`, and `shuffle` are never read [INFO] [stdout] --> src/nn/dataloader.rs:10:5 [INFO] [stdout] | [INFO] [stdout] 8 | pub struct DataLoader { [INFO] [stdout] | ---------- fields in this struct [INFO] [stdout] 9 | /// Dataset [INFO] [stdout] 10 | data: D, [INFO] [stdout] | ^^^^ [INFO] [stdout] 11 | /// Batch size [INFO] [stdout] 12 | batch_size: usize, [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] 13 | /// Whether or not to shuffle data [INFO] [stdout] 14 | shuffle: bool, [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `DataLoader` has a derived impl for the trait `Clone`, but this is intentionally ignored during dead code analysis [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: associated items `new` and `iter` are never used [INFO] [stdout] --> src/nn/dataloader.rs:41:8 [INFO] [stdout] | [INFO] [stdout] 33 | impl DataLoader { [INFO] [stdout] | -------------------------------------- associated items in this implementation [INFO] [stdout] ... [INFO] [stdout] 41 | fn new(data: D, batch_size: usize, shuffle: bool) -> DataLoader { [INFO] [stdout] | ^^^ [INFO] [stdout] ... [INFO] [stdout] 56 | fn iter(&mut self) -> DataLoader { [INFO] [stdout] | ^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: fields `img_dir`, `transform`, and `target_transform` are never read [INFO] [stdout] --> src/nn/dataset.rs:33:5 [INFO] [stdout] | [INFO] [stdout] 31 | pub struct CustomImageDataset { [INFO] [stdout] | ------------------ fields in this struct [INFO] [stdout] 32 | img_labels: Vec, [INFO] [stdout] 33 | img_dir: &'static str, [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] 34 | transform: Option, [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] 35 | target_transform: Option, [INFO] [stdout] | ^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: multiple fields are never read [INFO] [stdout] --> src/nn/optimizer.rs:38:5 [INFO] [stdout] | [INFO] [stdout] 36 | pub struct Adam { [INFO] [stdout] | ---- fields in this struct [INFO] [stdout] 37 | /// Learning rate [INFO] [stdout] 38 | lr: T, [INFO] [stdout] | ^^ [INFO] [stdout] 39 | /// Momentum [INFO] [stdout] 40 | momentum: f32, [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] 41 | /// Decay rate [INFO] [stdout] 42 | decay_rate: f32, [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] 43 | /// Beta 1 [INFO] [stdout] 44 | beta1: f32, [INFO] [stdout] | ^^^^^ [INFO] [stdout] 45 | /// Beta 2: [INFO] [stdout] 46 | beta2: f32, [INFO] [stdout] | ^^^^^ [INFO] [stdout] 47 | /// Epsilon [INFO] [stdout] 48 | epsilon: f32, [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] 49 | m_dw: f32, [INFO] [stdout] | ^^^^ [INFO] [stdout] 50 | v_dw: f32, [INFO] [stdout] | ^^^^ [INFO] [stdout] 51 | m_db: f32, [INFO] [stdout] | ^^^^ [INFO] [stdout] 52 | v_db: f32, [INFO] [stdout] | ^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: fields `lr`, `momentum`, and `decay_rate` are never read [INFO] [stdout] --> src/nn/optimizer.rs:97:5 [INFO] [stdout] | [INFO] [stdout] 95 | pub struct SGD { [INFO] [stdout] | --- fields in this struct [INFO] [stdout] 96 | /// Learning rate [INFO] [stdout] 97 | lr: T, [INFO] [stdout] | ^^ [INFO] [stdout] 98 | /// Momentum [INFO] [stdout] 99 | momentum: f32, [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] 100 | /// Decay rate [INFO] [stdout] 101 | decay_rate: f32, [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `swap` is never used [INFO] [stdout] --> src/tensor/mod.rs:33:4 [INFO] [stdout] | [INFO] [stdout] 33 | fn swap(lhs: &mut usize, rhs: &mut usize) { [INFO] [stdout] | ^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: methods `get_sub_tensor`, `sum`, and `prod` are never used [INFO] [stdout] --> src/tensor/mod.rs:686:8 [INFO] [stdout] | [INFO] [stdout] 560 | / impl<'a, T> Tensor<'a, T> [INFO] [stdout] 561 | | where [INFO] [stdout] 562 | | T: TensorElement + Div + Sum, [INFO] [stdout] 563 | | ::Err: Error + 'static, [INFO] [stdout] 564 | | Vec: IntoParallelIterator, [INFO] [stdout] 565 | | Vec<&'a T>: IntoParallelRefIterator<'a>, [INFO] [stdout] | |____________________________________________- methods in this implementation [INFO] [stdout] ... [INFO] [stdout] 686 | fn get_sub_tensor(&self, start_idx: Shape, size: Shape) -> Vec { [INFO] [stdout] | ^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 889 | fn sum(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^ [INFO] [stdout] ... [INFO] [stdout] 904 | fn prod(&self, rowcol: usize, dimension: Dimension) -> T { [INFO] [stdout] | ^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `ReLU` should have a snake case name [INFO] [stdout] --> src/nn/activation.rs:25:8 [INFO] [stdout] | [INFO] [stdout] 25 | pub fn ReLU<'a, 'b, T>(x: &Tensor<'a, T>) -> Tensor<'b, T> [INFO] [stdout] | ^^^^ help: convert the identifier to snake case: `re_lu` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(non_snake_case)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `PReLU` should have a snake case name [INFO] [stdout] --> src/nn/activation.rs:52:8 [INFO] [stdout] | [INFO] [stdout] 52 | pub fn PReLU<'a, 'b, T>(x: &Tensor<'a, T>, alpha: T) -> Tensor<'b, T> [INFO] [stdout] | ^^^^^ help: convert the identifier to snake case: `pre_lu` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `Sigmoid` should have a snake case name [INFO] [stdout] --> src/nn/activation.rs:79:8 [INFO] [stdout] | [INFO] [stdout] 79 | pub fn Sigmoid<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T> [INFO] [stdout] | ^^^^^^^ help: convert the identifier to snake case (notice the capitalization): `sigmoid` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `GeLU` should have a snake case name [INFO] [stdout] --> src/nn/activation.rs:108:8 [INFO] [stdout] | [INFO] [stdout] 108 | pub fn GeLU<'a, T>(x: &Tensor<'a, T>) -> Tensor<'a, T> [INFO] [stdout] | ^^^^ help: convert the identifier to snake case: `ge_lu` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: missing documentation for an associated function [INFO] [stdout] --> src/nn/loss.rs:14:5 [INFO] [stdout] | [INFO] [stdout] 14 | fn loss(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] note: the lint level is defined here [INFO] [stdout] --> src/nn/loss.rs:2:9 [INFO] [stdout] | [INFO] [stdout] 2 | #![warn(missing_docs)] [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: missing documentation for an associated function [INFO] [stdout] --> src/nn/loss.rs:15:5 [INFO] [stdout] | [INFO] [stdout] 15 | fn backward(); [INFO] [stdout] | ^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `BCEntroypyLoss` should have a snake case name [INFO] [stdout] --> src/nn/loss.rs:31:8 [INFO] [stdout] | [INFO] [stdout] 31 | pub fn BCEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^^^^^^^^^^ help: convert the identifier to snake case: `bcentroypy_loss` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `CEntroypyLoss` should have a snake case name [INFO] [stdout] --> src/nn/loss.rs:50:8 [INFO] [stdout] | [INFO] [stdout] 50 | pub fn CEntroypyLoss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^^^^^^^^^ help: convert the identifier to snake case: `centroypy_loss` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `L1Loss` should have a snake case name [INFO] [stdout] --> src/nn/loss.rs:73:8 [INFO] [stdout] | [INFO] [stdout] 73 | pub fn L1Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^^ help: convert the identifier to snake case: `l1_loss` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `L2Loss` should have a snake case name [INFO] [stdout] --> src/nn/loss.rs:96:8 [INFO] [stdout] | [INFO] [stdout] 96 | pub fn L2Loss<'a, T>(y: &Tensor<'a, T>, y_hat: &Tensor<'a, T>) -> T [INFO] [stdout] | ^^^^^^ help: convert the identifier to snake case: `l2_loss` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `MaxPool` should have a snake case name [INFO] [stdout] --> src/nn/pooling.rs:23:8 [INFO] [stdout] | [INFO] [stdout] 23 | pub fn MaxPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T> [INFO] [stdout] | ^^^^^^^ help: convert the identifier to snake case: `max_pool` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `MinPool` should have a snake case name [INFO] [stdout] --> src/nn/pooling.rs:62:8 [INFO] [stdout] | [INFO] [stdout] 62 | pub fn MinPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T> [INFO] [stdout] | ^^^^^^^ help: convert the identifier to snake case: `min_pool` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `AvgPool` should have a snake case name [INFO] [stdout] --> src/nn/pooling.rs:93:8 [INFO] [stdout] | [INFO] [stdout] 93 | pub fn AvgPool<'a, T>(x: &Tensor<'a, T>, stride: usize, padding: usize) -> Tensor<'a, T> [INFO] [stdout] | ^^^^^^^ help: convert the identifier to snake case: `avg_pool` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: missing documentation for a struct [INFO] [stdout] --> src/nn/transform.rs:4:1 [INFO] [stdout] | [INFO] [stdout] 4 | pub struct Transform; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] note: the lint level is defined here [INFO] [stdout] --> src/nn/transform.rs:2:9 [INFO] [stdout] | [INFO] [stdout] 2 | #![warn(missing_docs)] [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: missing documentation for a struct [INFO] [stdout] --> src/tensor/mod.rs:145:1 [INFO] [stdout] | [INFO] [stdout] 145 | pub struct Tensor<'a, T> [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] note: the lint level is defined here [INFO] [stdout] --> src/lib.rs:1:9 [INFO] [stdout] | [INFO] [stdout] 1 | #![warn(missing_docs)] [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] running `Command { std: "docker" "inspect" "61f0b901987c926f35d5ed5b191c927f8576776347465c5cf242ad49967a46d5", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "61f0b901987c926f35d5ed5b191c927f8576776347465c5cf242ad49967a46d5", kill_on_drop: false }` [INFO] [stdout] 61f0b901987c926f35d5ed5b191c927f8576776347465c5cf242ad49967a46d5