[INFO] cloning repository https://github.com/miny342/toy-deeplearning
[INFO] running `Command { std: "git" "-c" "credential.helper=" "-c" "credential.helper=/workspace/cargo-home/bin/git-credential-null" "clone" "--bare" "https://github.com/miny342/toy-deeplearning" "/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2Fminy342%2Ftoy-deeplearning", kill_on_drop: false }`
[INFO] [stderr] Cloning into bare repository '/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2Fminy342%2Ftoy-deeplearning'...
[INFO] running `Command { std: "git" "rev-parse" "HEAD", kill_on_drop: false }`
[INFO] [stdout] 32822f358a25cb555633e95e98e87e426f48c405
[INFO] checking miny342/toy-deeplearning against master#36b21637e93b038453924d3c66821089e71d8baa for pr-143164
[INFO] running `Command { std: "git" "clone" "/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2Fminy342%2Ftoy-deeplearning" "/workspace/builds/worker-6-tc1/source", kill_on_drop: false }`
[INFO] [stderr] Cloning into '/workspace/builds/worker-6-tc1/source'...
[INFO] [stderr] done.
[INFO] started tweaking git repo https://github.com/miny342/toy-deeplearning
[INFO] finished tweaking git repo https://github.com/miny342/toy-deeplearning
[INFO] tweaked toml for git repo https://github.com/miny342/toy-deeplearning written to /workspace/builds/worker-6-tc1/source/Cargo.toml
[INFO] validating manifest of git repo https://github.com/miny342/toy-deeplearning on toolchain 36b21637e93b038453924d3c66821089e71d8baa
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+36b21637e93b038453924d3c66821089e71d8baa" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }`
[INFO] crate git repo https://github.com/miny342/toy-deeplearning already has a lockfile, it will not be regenerated
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+36b21637e93b038453924d3c66821089e71d8baa" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-6-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-6-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:cf8efcab8866f2cf4285301c7418e2f4f2a9b088c91ba69c40d5b659f81557f7" "/opt/rustwide/cargo-home/bin/cargo" "+36b21637e93b038453924d3c66821089e71d8baa" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }`
[INFO] [stdout] e4f9d479bad47d2c273caf87b44b01d1e9d5450cd4422ed99e13e21289659d5e
[INFO] running `Command { std: "docker" "start" "-a" "e4f9d479bad47d2c273caf87b44b01d1e9d5450cd4422ed99e13e21289659d5e", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "inspect" "e4f9d479bad47d2c273caf87b44b01d1e9d5450cd4422ed99e13e21289659d5e", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "e4f9d479bad47d2c273caf87b44b01d1e9d5450cd4422ed99e13e21289659d5e", kill_on_drop: false }`
[INFO] [stdout] e4f9d479bad47d2c273caf87b44b01d1e9d5450cd4422ed99e13e21289659d5e
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-6-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-6-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:cf8efcab8866f2cf4285301c7418e2f4f2a9b088c91ba69c40d5b659f81557f7" "/opt/rustwide/cargo-home/bin/cargo" "+36b21637e93b038453924d3c66821089e71d8baa" "check" "--frozen" "--all" "--all-targets" "--message-format=json", kill_on_drop: false }`
[INFO] [stdout] 02d227e1342a8817db2d440d6d447509ec6c7bf7acbe230af9851834802b205d
[INFO] running `Command { std: "docker" "start" "-a" "02d227e1342a8817db2d440d6d447509ec6c7bf7acbe230af9851834802b205d", kill_on_drop: false }`
[INFO] [stderr]    Compiling libc v0.2.153
[INFO] [stderr]     Checking cfg-if v1.0.0
[INFO] [stderr]    Compiling libm v0.2.8
[INFO] [stderr]    Compiling autocfg v1.2.0
[INFO] [stderr]     Checking ppv-lite86 v0.2.17
[INFO] [stderr]    Compiling num-traits v0.2.18
[INFO] [stderr]     Checking getrandom v0.2.14
[INFO] [stderr]     Checking rand_core v0.6.4
[INFO] [stderr]     Checking rand_chacha v0.3.1
[INFO] [stderr]     Checking rand v0.8.5
[INFO] [stderr]     Checking rand_distr v0.4.3
[INFO] [stderr]     Checking toy-deeplearning v0.1.0 (/opt/rustwide/workdir)
[INFO] [stdout] warning: unused variable: `learning_rate`
[INFO] [stdout]   --> src/nn2.rs:27:26
[INFO] [stdout]    |
[INFO] [stdout] 27 |     fn update(&mut self, learning_rate: f64) {}
[INFO] [stdout]    |                          ^^^^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_learning_rate`
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(unused_variables)]` on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `learning_rate`
[INFO] [stdout]   --> src/nn2.rs:27:26
[INFO] [stdout]    |
[INFO] [stdout] 27 |     fn update(&mut self, learning_rate: f64) {}
[INFO] [stdout]    |                          ^^^^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_learning_rate`
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(unused_variables)]` on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `l`
[INFO] [stdout]   --> src/main.rs:67:9
[INFO] [stdout]    |
[INFO] [stdout] 67 |     for l in 0..1 {
[INFO] [stdout]    |         ^ help: if this is intentional, prefix it with an underscore: `_l`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `l`
[INFO] [stdout]   --> src/main.rs:67:9
[INFO] [stdout]    |
[INFO] [stdout] 67 |     for l in 0..1 {
[INFO] [stdout]    |         ^ help: if this is intentional, prefix it with an underscore: `_l`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated items `zeroed`, `normal`, `attach`, and `sum` are never used
[INFO] [stdout]   --> src/nn.rs:22:8
[INFO] [stdout]    |
[INFO] [stdout] 21 | impl<const R: usize, const C: usize> Matrix<R, C> {
[INFO] [stdout]    | ------------------------------------------------- associated items in this implementation
[INFO] [stdout] 22 |     fn zeroed() -> Self {
[INFO] [stdout]    |        ^^^^^^
[INFO] [stdout] ...
[INFO] [stdout] 25 |     fn normal() -> Self {
[INFO] [stdout]    |        ^^^^^^
[INFO] [stdout] ...
[INFO] [stdout] 30 |     fn attach<F: FnMut(f64) -> f64>(&self, mut f: F) -> Matrix<R, C> {
[INFO] [stdout]    |        ^^^^^^
[INFO] [stdout] ...
[INFO] [stdout] 59 |     fn sum(&self) -> f64 {
[INFO] [stdout]    |        ^^^
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(dead_code)]` on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: trait `Layer` is never used
[INFO] [stdout]    --> src/nn.rs:136:7
[INFO] [stdout]     |
[INFO] [stdout] 136 | trait Layer<const R: usize, const C: usize, const OR: usize = R, const OC: usize = C> {
[INFO] [stdout]     |       ^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: struct `LeakyRelu` is never constructed
[INFO] [stdout]    --> src/nn.rs:141:8
[INFO] [stdout]     |
[INFO] [stdout] 141 | struct LeakyRelu<const R: usize, const C: usize> {
[INFO] [stdout]     |        ^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `new` is never used
[INFO] [stdout]    --> src/nn.rs:146:8
[INFO] [stdout]     |
[INFO] [stdout] 145 | impl<const R: usize, const C: usize> LeakyRelu<R, C> {
[INFO] [stdout]     | ---------------------------------------------------- associated function in this implementation
[INFO] [stdout] 146 |     fn new() -> Self {
[INFO] [stdout]     |        ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: struct `Sigmoid` is never constructed
[INFO] [stdout]    --> src/nn.rs:161:8
[INFO] [stdout]     |
[INFO] [stdout] 161 | struct Sigmoid<const R: usize, const C: usize> {
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `new` is never used
[INFO] [stdout]    --> src/nn.rs:166:8
[INFO] [stdout]     |
[INFO] [stdout] 165 | impl<const R: usize, const C: usize> Sigmoid<R, C> {
[INFO] [stdout]     | -------------------------------------------------- associated function in this implementation
[INFO] [stdout] 166 |     fn new() -> Self {
[INFO] [stdout]     |        ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: struct `Affine` is never constructed
[INFO] [stdout]    --> src/nn.rs:181:8
[INFO] [stdout]     |
[INFO] [stdout] 181 | struct Affine<const I: usize, const O: usize, const B: usize> {
[INFO] [stdout]     |        ^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `new` is never used
[INFO] [stdout]    --> src/nn.rs:190:8
[INFO] [stdout]     |
[INFO] [stdout] 189 | impl<const I: usize, const O: usize, const B: usize> Affine<I, O, B> {
[INFO] [stdout]     | -------------------------------------------------------------------- associated function in this implementation
[INFO] [stdout] 190 |     fn new(w: Matrix<I, O>, b: Matrix<1, O>) -> Self {
[INFO] [stdout]     |        ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `softmax` is never used
[INFO] [stdout]    --> src/nn.rs:218:4
[INFO] [stdout]     |
[INFO] [stdout] 218 | fn softmax<const I: usize, const B: usize>(m: Matrix<B, I>) -> Matrix<B, I> {
[INFO] [stdout]     |    ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `cross_entropy_error` is never used
[INFO] [stdout]    --> src/nn.rs:248:4
[INFO] [stdout]     |
[INFO] [stdout] 248 | fn cross_entropy_error<const I: usize, const B: usize>(y: Matrix<B, I>, t: Matrix<B, I>) -> Matrix<B, 1> {
[INFO] [stdout]     |    ^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: struct `SoftmaxWithLoss` is never constructed
[INFO] [stdout]    --> src/nn.rs:261:8
[INFO] [stdout]     |
[INFO] [stdout] 261 | struct SoftmaxWithLoss<const I: usize, const B: usize> {
[INFO] [stdout]     |        ^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated items `new`, `forward`, and `backward` are never used
[INFO] [stdout]    --> src/nn.rs:268:8
[INFO] [stdout]     |
[INFO] [stdout] 267 | impl<const I: usize, const B: usize> SoftmaxWithLoss<I, B> {
[INFO] [stdout]     | ---------------------------------------------------------- associated items in this implementation
[INFO] [stdout] 268 |     fn new() -> Self {
[INFO] [stdout]     |        ^^^
[INFO] [stdout] ...
[INFO] [stdout] 271 |     fn forward(&mut self, x: Matrix<B, I>, t: Matrix<B, I>) -> Matrix<B, 1> {
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout] 277 |     fn backward(&self) -> Matrix<B, I> {
[INFO] [stdout]     |        ^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: struct `TwoLayerNet` is never constructed
[INFO] [stdout]    --> src/nn.rs:282:12
[INFO] [stdout]     |
[INFO] [stdout] 282 | pub struct TwoLayerNet<const I: usize, const H: usize, const O: usize, const B: usize> {
[INFO] [stdout]     |            ^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated items `zeroed`, `normal`, `attach`, and `sum` are never used
[INFO] [stdout]   --> src/nn.rs:22:8
[INFO] [stdout]    |
[INFO] [stdout] 21 | impl<const R: usize, const C: usize> Matrix<R, C> {
[INFO] [stdout]    | ------------------------------------------------- associated items in this implementation
[INFO] [stdout] 22 |     fn zeroed() -> Self {
[INFO] [stdout]    |        ^^^^^^
[INFO] [stdout] ...
[INFO] [stdout] 25 |     fn normal() -> Self {
[INFO] [stdout]    |        ^^^^^^
[INFO] [stdout] ...
[INFO] [stdout] 30 |     fn attach<F: FnMut(f64) -> f64>(&self, mut f: F) -> Matrix<R, C> {
[INFO] [stdout]    |        ^^^^^^
[INFO] [stdout] ...
[INFO] [stdout] 59 |     fn sum(&self) -> f64 {
[INFO] [stdout]    |        ^^^
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(dead_code)]` on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: trait `Layer` is never used
[INFO] [stdout]    --> src/nn.rs:136:7
[INFO] [stdout]     |
[INFO] [stdout] 136 | trait Layer<const R: usize, const C: usize, const OR: usize = R, const OC: usize = C> {
[INFO] [stdout]     |       ^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: struct `LeakyRelu` is never constructed
[INFO] [stdout]    --> src/nn.rs:141:8
[INFO] [stdout]     |
[INFO] [stdout] 141 | struct LeakyRelu<const R: usize, const C: usize> {
[INFO] [stdout]     |        ^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `new` is never used
[INFO] [stdout]    --> src/nn.rs:146:8
[INFO] [stdout]     |
[INFO] [stdout] 145 | impl<const R: usize, const C: usize> LeakyRelu<R, C> {
[INFO] [stdout]     | ---------------------------------------------------- associated function in this implementation
[INFO] [stdout] 146 |     fn new() -> Self {
[INFO] [stdout]     |        ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: struct `Sigmoid` is never constructed
[INFO] [stdout]    --> src/nn.rs:161:8
[INFO] [stdout]     |
[INFO] [stdout] 161 | struct Sigmoid<const R: usize, const C: usize> {
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated items `new`, `predict`, `loss`, `accuracy`, and `train` are never used
[INFO] [stdout]    --> src/nn.rs:290:12
[INFO] [stdout]     |
[INFO] [stdout] 289 | impl<const I: usize, const H: usize, const O: usize, const B: usize> TwoLayerNet<I, H, O, B> {
[INFO] [stdout]     | -------------------------------------------------------------------------------------------- associated items in this implementation
[INFO] [stdout] 290 |     pub fn new() -> Self {
[INFO] [stdout]     |            ^^^
[INFO] [stdout] ...
[INFO] [stdout] 298 |     pub fn predict(&mut self, x: Matrix<B, I>) -> Matrix<B, O> {
[INFO] [stdout]     |            ^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout] 304 |     pub fn loss(&mut self, x: Matrix<B, I>, t: Matrix<B, O>) -> Matrix<B, 1> {
[INFO] [stdout]     |            ^^^^
[INFO] [stdout] ...
[INFO] [stdout] 308 |     pub fn accuracy(&mut self, x: Matrix<B, I>, t: Matrix<B, O>) -> f64 {
[INFO] [stdout]     |            ^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout] 340 |     pub fn train(&mut self, x: Matrix<B, I>, t: Matrix<B, O>, learning_rate: f64) {
[INFO] [stdout]     |            ^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `new` is never used
[INFO] [stdout]    --> src/nn.rs:166:8
[INFO] [stdout]     |
[INFO] [stdout] 165 | impl<const R: usize, const C: usize> Sigmoid<R, C> {
[INFO] [stdout]     | -------------------------------------------------- associated function in this implementation
[INFO] [stdout] 166 |     fn new() -> Self {
[INFO] [stdout]     |        ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: struct `Affine` is never constructed
[INFO] [stdout]    --> src/nn.rs:181:8
[INFO] [stdout]     |
[INFO] [stdout] 181 | struct Affine<const I: usize, const O: usize, const B: usize> {
[INFO] [stdout]     |        ^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `new_onedim` is never used
[INFO] [stdout]   --> src/matrix.rs:93:12
[INFO] [stdout]    |
[INFO] [stdout] 79 | impl Matrix {
[INFO] [stdout]    | ----------- associated function in this implementation
[INFO] [stdout] ...
[INFO] [stdout] 93 |     pub fn new_onedim(size: (usize, usize), it: &[f64]) -> Matrix {
[INFO] [stdout]    |            ^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `new` is never used
[INFO] [stdout]    --> src/nn.rs:190:8
[INFO] [stdout]     |
[INFO] [stdout] 189 | impl<const I: usize, const O: usize, const B: usize> Affine<I, O, B> {
[INFO] [stdout]     | -------------------------------------------------------------------- associated function in this implementation
[INFO] [stdout] 190 |     fn new(w: Matrix<I, O>, b: Matrix<1, O>) -> Self {
[INFO] [stdout]     |        ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `softmax` is never used
[INFO] [stdout]    --> src/nn.rs:218:4
[INFO] [stdout]     |
[INFO] [stdout] 218 | fn softmax<const I: usize, const B: usize>(m: Matrix<B, I>) -> Matrix<B, I> {
[INFO] [stdout]     |    ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `new` is never used
[INFO] [stdout]   --> src/nn2.rs:39:8
[INFO] [stdout]    |
[INFO] [stdout] 38 | impl Affine {
[INFO] [stdout]    | ----------- associated function in this implementation
[INFO] [stdout] 39 |     fn new(w: Matrix, b: Matrix) -> Self {
[INFO] [stdout]    |        ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `cross_entropy_error` is never used
[INFO] [stdout]    --> src/nn.rs:248:4
[INFO] [stdout]     |
[INFO] [stdout] 248 | fn cross_entropy_error<const I: usize, const B: usize>(y: Matrix<B, I>, t: Matrix<B, I>) -> Matrix<B, 1> {
[INFO] [stdout]     |    ^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: struct `SoftmaxWithLoss` is never constructed
[INFO] [stdout]    --> src/nn.rs:261:8
[INFO] [stdout]     |
[INFO] [stdout] 261 | struct SoftmaxWithLoss<const I: usize, const B: usize> {
[INFO] [stdout]     |        ^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated items `new`, `forward`, and `backward` are never used
[INFO] [stdout]    --> src/nn.rs:268:8
[INFO] [stdout]     |
[INFO] [stdout] 267 | impl<const I: usize, const B: usize> SoftmaxWithLoss<I, B> {
[INFO] [stdout]     | ---------------------------------------------------------- associated items in this implementation
[INFO] [stdout] 268 |     fn new() -> Self {
[INFO] [stdout]     |        ^^^
[INFO] [stdout] ...
[INFO] [stdout] 271 |     fn forward(&mut self, x: Matrix<B, I>, t: Matrix<B, I>) -> Matrix<B, 1> {
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout] 277 |     fn backward(&self) -> Matrix<B, I> {
[INFO] [stdout]     |        ^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: struct `TwoLayerNet` is never constructed
[INFO] [stdout]    --> src/nn.rs:282:12
[INFO] [stdout]     |
[INFO] [stdout] 282 | pub struct TwoLayerNet<const I: usize, const H: usize, const O: usize, const B: usize> {
[INFO] [stdout]     |            ^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated items `new`, `predict`, `loss`, `accuracy`, and `train` are never used
[INFO] [stdout]    --> src/nn.rs:290:12
[INFO] [stdout]     |
[INFO] [stdout] 289 | impl<const I: usize, const H: usize, const O: usize, const B: usize> TwoLayerNet<I, H, O, B> {
[INFO] [stdout]     | -------------------------------------------------------------------------------------------- associated items in this implementation
[INFO] [stdout] 290 |     pub fn new() -> Self {
[INFO] [stdout]     |            ^^^
[INFO] [stdout] ...
[INFO] [stdout] 298 |     pub fn predict(&mut self, x: Matrix<B, I>) -> Matrix<B, O> {
[INFO] [stdout]     |            ^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout] 304 |     pub fn loss(&mut self, x: Matrix<B, I>, t: Matrix<B, O>) -> Matrix<B, 1> {
[INFO] [stdout]     |            ^^^^
[INFO] [stdout] ...
[INFO] [stdout] 308 |     pub fn accuracy(&mut self, x: Matrix<B, I>, t: Matrix<B, O>) -> f64 {
[INFO] [stdout]     |            ^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout] 340 |     pub fn train(&mut self, x: Matrix<B, I>, t: Matrix<B, O>, learning_rate: f64) {
[INFO] [stdout]     |            ^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `new_onedim` is never used
[INFO] [stdout]   --> src/matrix.rs:93:12
[INFO] [stdout]    |
[INFO] [stdout] 79 | impl Matrix {
[INFO] [stdout]    | ----------- associated function in this implementation
[INFO] [stdout] ...
[INFO] [stdout] 93 |     pub fn new_onedim(size: (usize, usize), it: &[f64]) -> Matrix {
[INFO] [stdout]    |            ^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `new` is never used
[INFO] [stdout]   --> src/nn2.rs:39:8
[INFO] [stdout]    |
[INFO] [stdout] 38 | impl Affine {
[INFO] [stdout]    | ----------- associated function in this implementation
[INFO] [stdout] 39 |     fn new(w: Matrix, b: Matrix) -> Self {
[INFO] [stdout]    |        ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stderr]     Finished `dev` profile [unoptimized + debuginfo] target(s) in 2.17s
[INFO] running `Command { std: "docker" "inspect" "02d227e1342a8817db2d440d6d447509ec6c7bf7acbe230af9851834802b205d", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "02d227e1342a8817db2d440d6d447509ec6c7bf7acbe230af9851834802b205d", kill_on_drop: false }`
[INFO] [stdout] 02d227e1342a8817db2d440d6d447509ec6c7bf7acbe230af9851834802b205d
