[INFO] cloning repository https://github.com/ano333333/zero-deeplearning
[INFO] running `Command { std: "git" "-c" "credential.helper=" "-c" "credential.helper=/workspace/cargo-home/bin/git-credential-null" "clone" "--bare" "https://github.com/ano333333/zero-deeplearning" "/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2Fano333333%2Fzero-deeplearning", kill_on_drop: false }`
[INFO] [stderr] Cloning into bare repository '/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2Fano333333%2Fzero-deeplearning'...
[INFO] running `Command { std: "git" "rev-parse" "HEAD", kill_on_drop: false }`
[INFO] [stdout] 1279e9bd43561db6bb25a6e26fcfac2f3c038558
[INFO] checking ano333333/zero-deeplearning against master#bca37a20bd376ce3fd138e7cdee7fe704e0f8814 for pr-139493-3
[INFO] running `Command { std: "git" "clone" "/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2Fano333333%2Fzero-deeplearning" "/workspace/builds/worker-3-tc1/source", kill_on_drop: false }`
[INFO] [stderr] Cloning into '/workspace/builds/worker-3-tc1/source'...
[INFO] [stderr] done.
[INFO] started tweaking git repo https://github.com/ano333333/zero-deeplearning
[INFO] finished tweaking git repo https://github.com/ano333333/zero-deeplearning
[INFO] tweaked toml for git repo https://github.com/ano333333/zero-deeplearning written to /workspace/builds/worker-3-tc1/source/Cargo.toml
[INFO] validating manifest of git repo https://github.com/ano333333/zero-deeplearning on toolchain bca37a20bd376ce3fd138e7cdee7fe704e0f8814
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+bca37a20bd376ce3fd138e7cdee7fe704e0f8814" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }`
[INFO] crate git repo https://github.com/ano333333/zero-deeplearning already has a lockfile, it will not be regenerated
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+bca37a20bd376ce3fd138e7cdee7fe704e0f8814" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }`
[INFO] [stderr]     Updating crates.io index
[INFO] [stderr]  Downloading crates ...
[INFO] [stderr]   Downloaded pathfinder_geometry v0.5.1
[INFO] [stderr]   Downloaded pathfinder_simd v0.5.3
[INFO] [stderr]   Downloaded cstr v0.2.12
[INFO] [stderr]   Downloaded float-ord v0.3.2
[INFO] [stderr]   Downloaded yeslogic-fontconfig-sys v5.0.0
[INFO] [stderr]   Downloaded mnist v0.6.0
[INFO] [stderr]   Downloaded ndarray-rand v0.14.0
[INFO] [stderr]   Downloaded plotters-backend v0.3.6
[INFO] [stderr]   Downloaded plotters-bitmap v0.3.6
[INFO] [stderr]   Downloaded plotters-svg v0.3.6
[INFO] [stderr]   Downloaded font-kit v0.13.2
[INFO] [stderr]   Downloaded plotters v0.3.6
[INFO] [stderr]   Downloaded freetype-sys v0.20.1
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-3-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-3-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:aa71247004a7fa38d13ec170f48f06cdedf5bc50b2a8645e56ed7e992e6fa513" "/opt/rustwide/cargo-home/bin/cargo" "+bca37a20bd376ce3fd138e7cdee7fe704e0f8814" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }`
[INFO] [stdout] 62449cba2b14a3b9dbc1a979094e5f882e91cea3c39d5ae5f86721ef342b514f
[INFO] running `Command { std: "docker" "start" "-a" "62449cba2b14a3b9dbc1a979094e5f882e91cea3c39d5ae5f86721ef342b514f", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "inspect" "62449cba2b14a3b9dbc1a979094e5f882e91cea3c39d5ae5f86721ef342b514f", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "62449cba2b14a3b9dbc1a979094e5f882e91cea3c39d5ae5f86721ef342b514f", kill_on_drop: false }`
[INFO] [stdout] 62449cba2b14a3b9dbc1a979094e5f882e91cea3c39d5ae5f86721ef342b514f
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-3-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-3-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:aa71247004a7fa38d13ec170f48f06cdedf5bc50b2a8645e56ed7e992e6fa513" "/opt/rustwide/cargo-home/bin/cargo" "+bca37a20bd376ce3fd138e7cdee7fe704e0f8814" "check" "--frozen" "--all" "--all-targets" "--message-format=json", kill_on_drop: false }`
[INFO] [stdout] b9c4e8843643e348ab92f0ace74247f38f0941a7f189e4d36c81e1d30617f9a8
[INFO] running `Command { std: "docker" "start" "-a" "b9c4e8843643e348ab92f0ace74247f38f0941a7f189e4d36c81e1d30617f9a8", kill_on_drop: false }`
[INFO] [stderr]    Compiling proc-macro2 v1.0.84
[INFO] [stderr]    Compiling num-traits v0.2.19
[INFO] [stderr]    Compiling yeslogic-fontconfig-sys v5.0.0
[INFO] [stderr]    Compiling freetype-sys v0.20.1
[INFO] [stderr]     Checking miniz_oxide v0.7.3
[INFO] [stderr]    Compiling rustc_version v0.4.0
[INFO] [stderr]    Compiling matrixmultiply v0.3.8
[INFO] [stderr]     Checking libloading v0.8.3
[INFO] [stderr]     Checking dirs-sys-next v0.1.2
[INFO] [stderr]     Checking dlib v0.5.2
[INFO] [stderr]     Checking plotters-backend v0.3.6
[INFO] [stderr]     Checking weezl v0.1.8
[INFO] [stderr]     Checking jpeg-decoder v0.3.1
[INFO] [stderr]     Checking bytemuck v1.16.0
[INFO] [stderr]    Compiling quote v1.0.36
[INFO] [stderr]    Compiling font-kit v0.13.2
[INFO] [stderr]    Compiling pathfinder_simd v0.5.3
[INFO] [stderr]     Checking flate2 v1.0.30
[INFO] [stderr]     Checking gif v0.12.0
[INFO] [stderr]     Checking dirs-next v2.0.0
[INFO] [stderr]    Compiling cstr v0.2.12
[INFO] [stderr]     Checking rand v0.8.5
[INFO] [stderr]     Checking float-ord v0.3.2
[INFO] [stderr]     Checking plotters-svg v0.3.6
[INFO] [stderr]     Checking png v0.17.13
[INFO] [stderr]     Checking ttf-parser v0.20.0
[INFO] [stderr]     Checking mnist v0.6.0
[INFO] [stderr]     Checking pathfinder_geometry v0.5.1
[INFO] [stderr]     Checking num-complex v0.4.6
[INFO] [stderr]     Checking num-integer v0.1.46
[INFO] [stderr]     Checking chrono v0.4.38
[INFO] [stderr]     Checking rand_distr v0.4.3
[INFO] [stderr]     Checking ndarray v0.15.6
[INFO] [stderr]     Checking image v0.24.9
[INFO] [stderr]     Checking plotters-bitmap v0.3.6
[INFO] [stderr]     Checking plotters v0.3.6
[INFO] [stderr]     Checking ndarray-rand v0.14.0
[INFO] [stderr]     Checking zero-deeplearning v0.1.0 (/opt/rustwide/workdir)
[INFO] [stdout] warning: unused import: `layer::affine_layer::AffineLayer`
[INFO] [stdout]  --> src/main.rs:2:5
[INFO] [stdout]   |
[INFO] [stdout] 2 | use layer::affine_layer::AffineLayer;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout]   = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `layer::layer::Layer`
[INFO] [stdout]  --> src/main.rs:3:5
[INFO] [stdout]   |
[INFO] [stdout] 3 | use layer::layer::Layer;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `layer::relu_layer::ReluLayer`
[INFO] [stdout]  --> src/main.rs:4:5
[INFO] [stdout]   |
[INFO] [stdout] 4 | use layer::relu_layer::ReluLayer;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `layer::softmax_with_loss_layer::SoftmaxWithLossLayer`
[INFO] [stdout]  --> src/main.rs:5:5
[INFO] [stdout]   |
[INFO] [stdout] 5 | use layer::softmax_with_loss_layer::SoftmaxWithLossLayer;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `two_layer_net::TwoLayerNetGradient`
[INFO] [stdout]   --> src/main.rs:17:5
[INFO] [stdout]    |
[INFO] [stdout] 17 | use two_layer_net::TwoLayerNetGradient;
[INFO] [stdout]    |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::layer::batch_normalization_layer::BatchNormalizationLayer`
[INFO] [stdout]   --> src/main.rs:19:5
[INFO] [stdout]    |
[INFO] [stdout] 19 | use crate::layer::batch_normalization_layer::BatchNormalizationLayer;
[INFO] [stdout]    |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `layer::affine_layer::AffineLayer`
[INFO] [stdout]  --> src/main.rs:2:5
[INFO] [stdout]   |
[INFO] [stdout] 2 | use layer::affine_layer::AffineLayer;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout]   = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `layer::layer::Layer`
[INFO] [stdout]  --> src/main.rs:3:5
[INFO] [stdout]   |
[INFO] [stdout] 3 | use layer::layer::Layer;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `layer::relu_layer::ReluLayer`
[INFO] [stdout]  --> src/main.rs:4:5
[INFO] [stdout]   |
[INFO] [stdout] 4 | use layer::relu_layer::ReluLayer;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `layer::softmax_with_loss_layer::SoftmaxWithLossLayer`
[INFO] [stdout]  --> src/main.rs:5:5
[INFO] [stdout]   |
[INFO] [stdout] 5 | use layer::softmax_with_loss_layer::SoftmaxWithLossLayer;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `two_layer_net::TwoLayerNetGradient`
[INFO] [stdout]   --> src/main.rs:17:5
[INFO] [stdout]    |
[INFO] [stdout] 17 | use two_layer_net::TwoLayerNetGradient;
[INFO] [stdout]    |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::layer::batch_normalization_layer::BatchNormalizationLayer`
[INFO] [stdout]   --> src/main.rs:19:5
[INFO] [stdout]    |
[INFO] [stdout] 19 | use crate::layer::batch_normalization_layer::BatchNormalizationLayer;
[INFO] [stdout]    |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `du2`
[INFO] [stdout]   --> src/layer/batch_normalization_layer.rs:59:13
[INFO] [stdout]    |
[INFO] [stdout] 59 |         let du2 = (&dxhat * &self.u1).sum_axis(Axis(1));
[INFO] [stdout]    |             ^^^ help: if this is intentional, prefix it with an underscore: `_du2`
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(unused_variables)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: struct `AddLayer` is never constructed
[INFO] [stdout]  --> src/layer/add_layer.rs:4:12
[INFO] [stdout]   |
[INFO] [stdout] 4 | pub struct AddLayer<Dim: Dimension> {
[INFO] [stdout]   |            ^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout]   = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `new` is never used
[INFO] [stdout]  --> src/layer/add_layer.rs:9:12
[INFO] [stdout]   |
[INFO] [stdout] 8 | impl<Dim: Dimension> AddLayer<Dim> {
[INFO] [stdout]   | ---------------------------------- associated function in this implementation
[INFO] [stdout] 9 |     pub fn new() -> Self {
[INFO] [stdout]   |            ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: struct `DivLayer` is never constructed
[INFO] [stdout]  --> src/layer/div_layer.rs:4:12
[INFO] [stdout]   |
[INFO] [stdout] 4 | pub struct DivLayer<Dim: Dimension> {
[INFO] [stdout]   |            ^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `new` is never used
[INFO] [stdout]  --> src/layer/div_layer.rs:9:12
[INFO] [stdout]   |
[INFO] [stdout] 8 | impl<Dim: Dimension> DivLayer<Dim> {
[INFO] [stdout]   | ---------------------------------- associated function in this implementation
[INFO] [stdout] 9 |     pub fn new() -> Self {
[INFO] [stdout]   |            ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: struct `ExpLayer` is never constructed
[INFO] [stdout]  --> src/layer/exp_layer.rs:4:12
[INFO] [stdout]   |
[INFO] [stdout] 4 | pub struct ExpLayer<Dim: Dimension> {
[INFO] [stdout]   |            ^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `new` is never used
[INFO] [stdout]  --> src/layer/exp_layer.rs:9:12
[INFO] [stdout]   |
[INFO] [stdout] 8 | impl<Dim: Dimension> ExpLayer<Dim> {
[INFO] [stdout]   | ---------------------------------- associated function in this implementation
[INFO] [stdout] 9 |     pub fn new() -> Self {
[INFO] [stdout]   |            ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: struct `MulLayer` is never constructed
[INFO] [stdout]  --> src/layer/mul_layer.rs:4:12
[INFO] [stdout]   |
[INFO] [stdout] 4 | pub struct MulLayer<Dim: Dimension> {
[INFO] [stdout]   |            ^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `new` is never used
[INFO] [stdout]   --> src/layer/mul_layer.rs:10:12
[INFO] [stdout]    |
[INFO] [stdout]  9 | impl<Dim: Dimension> MulLayer<Dim> {
[INFO] [stdout]    | ---------------------------------- associated function in this implementation
[INFO] [stdout] 10 |     pub fn new() -> Self {
[INFO] [stdout]    |            ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: struct `SigmoidLayer` is never constructed
[INFO] [stdout]  --> src/layer/sigmoid_layer.rs:4:12
[INFO] [stdout]   |
[INFO] [stdout] 4 | pub struct SigmoidLayer<Dim: Dimension> {
[INFO] [stdout]   |            ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `new` is never used
[INFO] [stdout]  --> src/layer/sigmoid_layer.rs:9:12
[INFO] [stdout]   |
[INFO] [stdout] 8 | impl<Dim: Dimension> SigmoidLayer<Dim> {
[INFO] [stdout]   | -------------------------------------- associated function in this implementation
[INFO] [stdout] 9 |     pub fn new() -> Self {
[INFO] [stdout]   |            ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: struct `AdaGrad` is never constructed
[INFO] [stdout]  --> src/optimize/ada_grad.rs:5:12
[INFO] [stdout]   |
[INFO] [stdout] 5 | pub struct AdaGrad<D: Dimension> {
[INFO] [stdout]   |            ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `new` is never used
[INFO] [stdout]   --> src/optimize/ada_grad.rs:11:12
[INFO] [stdout]    |
[INFO] [stdout] 10 | impl<D: Dimension> AdaGrad<D> {
[INFO] [stdout]    | ----------------------------- associated function in this implementation
[INFO] [stdout] 11 |     pub fn new(learning_rate: f64) -> Self {
[INFO] [stdout]    |            ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: struct `Momentum` is never constructed
[INFO] [stdout]  --> src/optimize/momentum.rs:5:12
[INFO] [stdout]   |
[INFO] [stdout] 5 | pub struct Momentum<D: Dimension> {
[INFO] [stdout]   |            ^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `new` is never used
[INFO] [stdout]   --> src/optimize/momentum.rs:12:12
[INFO] [stdout]    |
[INFO] [stdout] 11 | impl<D: Dimension> Momentum<D> {
[INFO] [stdout]    | ------------------------------ associated function in this implementation
[INFO] [stdout] 12 |     pub fn new(learning_rate: f64, momentum: f64) -> Self {
[INFO] [stdout]    |            ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `identity_function` is never used
[INFO] [stdout]  --> src/subfunction/identity_function.rs:3:8
[INFO] [stdout]   |
[INFO] [stdout] 3 | pub fn identity_function<D: Dimension>(x: ArrayView<f64, D>) -> Array<f64, D> {
[INFO] [stdout]   |        ^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `numerical_gradient` is never used
[INFO] [stdout]  --> src/subfunction/numerical_gradient.rs:3:8
[INFO] [stdout]   |
[INFO] [stdout] 3 | pub fn numerical_gradient<D: Dimension>(
[INFO] [stdout]   |        ^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `relu` is never used
[INFO] [stdout]  --> src/subfunction/relu.rs:3:8
[INFO] [stdout]   |
[INFO] [stdout] 3 | pub fn relu<D: Dimension>(x: ArrayView<f64, D>) -> Array<f64, D> {
[INFO] [stdout]   |        ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `sigmoid` is never used
[INFO] [stdout]  --> src/subfunction/sigmoid.rs:3:8
[INFO] [stdout]   |
[INFO] [stdout] 3 | pub fn sigmoid<D: Dimension>(x: ArrayView<f64, D>) -> Array<f64, D> {
[INFO] [stdout]   |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `softmax` is never used
[INFO] [stdout]  --> src/subfunction/softmax.rs:3:8
[INFO] [stdout]   |
[INFO] [stdout] 3 | pub fn softmax<D: Dimension>(x: ArrayView<f64, D>) -> Array<f64, D> {
[INFO] [stdout]   |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `step_function` is never used
[INFO] [stdout]  --> src/subfunction/step_function.rs:3:8
[INFO] [stdout]   |
[INFO] [stdout] 3 | pub fn step_function<D: Dimension>(x: ArrayView<f64, D>) -> Array<f64, D> {
[INFO] [stdout]   |        ^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: hiding a lifetime that's elided elsewhere is confusing
[INFO] [stdout]   --> src/two_layer_net.rs:48:27
[INFO] [stdout]    |
[INFO] [stdout] 48 |     pub fn create_affine1(&self) -> AffineLayer {
[INFO] [stdout]    |                           ^^^^^     ^^^^^^^^^^^ the same lifetime is hidden here
[INFO] [stdout]    |                           |
[INFO] [stdout]    |                           the lifetime is elided here
[INFO] [stdout]    |
[INFO] [stdout]    = help: the same lifetime is referred to in inconsistent ways, making the signature confusing
[INFO] [stdout]    = note: `#[warn(mismatched_lifetime_syntaxes)]` on by default
[INFO] [stdout] help: use `'_` for type paths
[INFO] [stdout]    |
[INFO] [stdout] 48 |     pub fn create_affine1(&self) -> AffineLayer<'_> {
[INFO] [stdout]    |                                                ++++
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: hiding a lifetime that's elided elsewhere is confusing
[INFO] [stdout]   --> src/two_layer_net.rs:51:40
[INFO] [stdout]    |
[INFO] [stdout] 51 |     pub fn create_batch_normalization1(&self) -> BatchNormalizationLayer {
[INFO] [stdout]    |                                        ^^^^^     ^^^^^^^^^^^^^^^^^^^^^^^ the same lifetime is hidden here
[INFO] [stdout]    |                                        |
[INFO] [stdout]    |                                        the lifetime is elided here
[INFO] [stdout]    |
[INFO] [stdout]    = help: the same lifetime is referred to in inconsistent ways, making the signature confusing
[INFO] [stdout] help: use `'_` for type paths
[INFO] [stdout]    |
[INFO] [stdout] 51 |     pub fn create_batch_normalization1(&self) -> BatchNormalizationLayer<'_> {
[INFO] [stdout]    |                                                                         ++++
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: hiding a lifetime that's elided elsewhere is confusing
[INFO] [stdout]   --> src/two_layer_net.rs:57:27
[INFO] [stdout]    |
[INFO] [stdout] 57 |     pub fn create_affine2(&self) -> AffineLayer {
[INFO] [stdout]    |                           ^^^^^     ^^^^^^^^^^^ the same lifetime is hidden here
[INFO] [stdout]    |                           |
[INFO] [stdout]    |                           the lifetime is elided here
[INFO] [stdout]    |
[INFO] [stdout]    = help: the same lifetime is referred to in inconsistent ways, making the signature confusing
[INFO] [stdout] help: use `'_` for type paths
[INFO] [stdout]    |
[INFO] [stdout] 57 |     pub fn create_affine2(&self) -> AffineLayer<'_> {
[INFO] [stdout]    |                                                ++++
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `du2`
[INFO] [stdout]   --> src/layer/batch_normalization_layer.rs:59:13
[INFO] [stdout]    |
[INFO] [stdout] 59 |         let du2 = (&dxhat * &self.u1).sum_axis(Axis(1));
[INFO] [stdout]    |             ^^^ help: if this is intentional, prefix it with an underscore: `_du2`
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(unused_variables)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: struct `AddLayer` is never constructed
[INFO] [stdout]  --> src/layer/add_layer.rs:4:12
[INFO] [stdout]   |
[INFO] [stdout] 4 | pub struct AddLayer<Dim: Dimension> {
[INFO] [stdout]   |            ^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout]   = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `new` is never used
[INFO] [stdout]  --> src/layer/add_layer.rs:9:12
[INFO] [stdout]   |
[INFO] [stdout] 8 | impl<Dim: Dimension> AddLayer<Dim> {
[INFO] [stdout]   | ---------------------------------- associated function in this implementation
[INFO] [stdout] 9 |     pub fn new() -> Self {
[INFO] [stdout]   |            ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: struct `DivLayer` is never constructed
[INFO] [stdout]  --> src/layer/div_layer.rs:4:12
[INFO] [stdout]   |
[INFO] [stdout] 4 | pub struct DivLayer<Dim: Dimension> {
[INFO] [stdout]   |            ^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `new` is never used
[INFO] [stdout]  --> src/layer/div_layer.rs:9:12
[INFO] [stdout]   |
[INFO] [stdout] 8 | impl<Dim: Dimension> DivLayer<Dim> {
[INFO] [stdout]   | ---------------------------------- associated function in this implementation
[INFO] [stdout] 9 |     pub fn new() -> Self {
[INFO] [stdout]   |            ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: struct `ExpLayer` is never constructed
[INFO] [stdout]  --> src/layer/exp_layer.rs:4:12
[INFO] [stdout]   |
[INFO] [stdout] 4 | pub struct ExpLayer<Dim: Dimension> {
[INFO] [stdout]   |            ^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `new` is never used
[INFO] [stdout]  --> src/layer/exp_layer.rs:9:12
[INFO] [stdout]   |
[INFO] [stdout] 8 | impl<Dim: Dimension> ExpLayer<Dim> {
[INFO] [stdout]   | ---------------------------------- associated function in this implementation
[INFO] [stdout] 9 |     pub fn new() -> Self {
[INFO] [stdout]   |            ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: struct `MulLayer` is never constructed
[INFO] [stdout]  --> src/layer/mul_layer.rs:4:12
[INFO] [stdout]   |
[INFO] [stdout] 4 | pub struct MulLayer<Dim: Dimension> {
[INFO] [stdout]   |            ^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `new` is never used
[INFO] [stdout]   --> src/layer/mul_layer.rs:10:12
[INFO] [stdout]    |
[INFO] [stdout]  9 | impl<Dim: Dimension> MulLayer<Dim> {
[INFO] [stdout]    | ---------------------------------- associated function in this implementation
[INFO] [stdout] 10 |     pub fn new() -> Self {
[INFO] [stdout]    |            ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: struct `SigmoidLayer` is never constructed
[INFO] [stdout]  --> src/layer/sigmoid_layer.rs:4:12
[INFO] [stdout]   |
[INFO] [stdout] 4 | pub struct SigmoidLayer<Dim: Dimension> {
[INFO] [stdout]   |            ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `new` is never used
[INFO] [stdout]  --> src/layer/sigmoid_layer.rs:9:12
[INFO] [stdout]   |
[INFO] [stdout] 8 | impl<Dim: Dimension> SigmoidLayer<Dim> {
[INFO] [stdout]   | -------------------------------------- associated function in this implementation
[INFO] [stdout] 9 |     pub fn new() -> Self {
[INFO] [stdout]   |            ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: struct `AdaGrad` is never constructed
[INFO] [stdout]  --> src/optimize/ada_grad.rs:5:12
[INFO] [stdout]   |
[INFO] [stdout] 5 | pub struct AdaGrad<D: Dimension> {
[INFO] [stdout]   |            ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `new` is never used
[INFO] [stdout]   --> src/optimize/ada_grad.rs:11:12
[INFO] [stdout]    |
[INFO] [stdout] 10 | impl<D: Dimension> AdaGrad<D> {
[INFO] [stdout]    | ----------------------------- associated function in this implementation
[INFO] [stdout] 11 |     pub fn new(learning_rate: f64) -> Self {
[INFO] [stdout]    |            ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: struct `Momentum` is never constructed
[INFO] [stdout]  --> src/optimize/momentum.rs:5:12
[INFO] [stdout]   |
[INFO] [stdout] 5 | pub struct Momentum<D: Dimension> {
[INFO] [stdout]   |            ^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `new` is never used
[INFO] [stdout]   --> src/optimize/momentum.rs:12:12
[INFO] [stdout]    |
[INFO] [stdout] 11 | impl<D: Dimension> Momentum<D> {
[INFO] [stdout]    | ------------------------------ associated function in this implementation
[INFO] [stdout] 12 |     pub fn new(learning_rate: f64, momentum: f64) -> Self {
[INFO] [stdout]    |            ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `identity_function` is never used
[INFO] [stdout]  --> src/subfunction/identity_function.rs:3:8
[INFO] [stdout]   |
[INFO] [stdout] 3 | pub fn identity_function<D: Dimension>(x: ArrayView<f64, D>) -> Array<f64, D> {
[INFO] [stdout]   |        ^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `numerical_gradient` is never used
[INFO] [stdout]  --> src/subfunction/numerical_gradient.rs:3:8
[INFO] [stdout]   |
[INFO] [stdout] 3 | pub fn numerical_gradient<D: Dimension>(
[INFO] [stdout]   |        ^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `relu` is never used
[INFO] [stdout]  --> src/subfunction/relu.rs:3:8
[INFO] [stdout]   |
[INFO] [stdout] 3 | pub fn relu<D: Dimension>(x: ArrayView<f64, D>) -> Array<f64, D> {
[INFO] [stdout]   |        ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `sigmoid` is never used
[INFO] [stdout]  --> src/subfunction/sigmoid.rs:3:8
[INFO] [stdout]   |
[INFO] [stdout] 3 | pub fn sigmoid<D: Dimension>(x: ArrayView<f64, D>) -> Array<f64, D> {
[INFO] [stdout]   |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `softmax` is never used
[INFO] [stdout]  --> src/subfunction/softmax.rs:3:8
[INFO] [stdout]   |
[INFO] [stdout] 3 | pub fn softmax<D: Dimension>(x: ArrayView<f64, D>) -> Array<f64, D> {
[INFO] [stdout]   |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `step_function` is never used
[INFO] [stdout]  --> src/subfunction/step_function.rs:3:8
[INFO] [stdout]   |
[INFO] [stdout] 3 | pub fn step_function<D: Dimension>(x: ArrayView<f64, D>) -> Array<f64, D> {
[INFO] [stdout]   |        ^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: hiding a lifetime that's elided elsewhere is confusing
[INFO] [stdout]   --> src/two_layer_net.rs:48:27
[INFO] [stdout]    |
[INFO] [stdout] 48 |     pub fn create_affine1(&self) -> AffineLayer {
[INFO] [stdout]    |                           ^^^^^     ^^^^^^^^^^^ the same lifetime is hidden here
[INFO] [stdout]    |                           |
[INFO] [stdout]    |                           the lifetime is elided here
[INFO] [stdout]    |
[INFO] [stdout]    = help: the same lifetime is referred to in inconsistent ways, making the signature confusing
[INFO] [stdout]    = note: `#[warn(mismatched_lifetime_syntaxes)]` on by default
[INFO] [stdout] help: use `'_` for type paths
[INFO] [stdout]    |
[INFO] [stdout] 48 |     pub fn create_affine1(&self) -> AffineLayer<'_> {
[INFO] [stdout]    |                                                ++++
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: hiding a lifetime that's elided elsewhere is confusing
[INFO] [stdout]   --> src/two_layer_net.rs:51:40
[INFO] [stdout]    |
[INFO] [stdout] 51 |     pub fn create_batch_normalization1(&self) -> BatchNormalizationLayer {
[INFO] [stdout]    |                                        ^^^^^     ^^^^^^^^^^^^^^^^^^^^^^^ the same lifetime is hidden here
[INFO] [stdout]    |                                        |
[INFO] [stdout]    |                                        the lifetime is elided here
[INFO] [stdout]    |
[INFO] [stdout]    = help: the same lifetime is referred to in inconsistent ways, making the signature confusing
[INFO] [stdout] help: use `'_` for type paths
[INFO] [stdout]    |
[INFO] [stdout] 51 |     pub fn create_batch_normalization1(&self) -> BatchNormalizationLayer<'_> {
[INFO] [stdout]    |                                                                         ++++
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: hiding a lifetime that's elided elsewhere is confusing
[INFO] [stdout]   --> src/two_layer_net.rs:57:27
[INFO] [stdout]    |
[INFO] [stdout] 57 |     pub fn create_affine2(&self) -> AffineLayer {
[INFO] [stdout]    |                           ^^^^^     ^^^^^^^^^^^ the same lifetime is hidden here
[INFO] [stdout]    |                           |
[INFO] [stdout]    |                           the lifetime is elided here
[INFO] [stdout]    |
[INFO] [stdout]    = help: the same lifetime is referred to in inconsistent ways, making the signature confusing
[INFO] [stdout] help: use `'_` for type paths
[INFO] [stdout]    |
[INFO] [stdout] 57 |     pub fn create_affine2(&self) -> AffineLayer<'_> {
[INFO] [stdout]    |                                                ++++
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stderr]     Finished `dev` profile [unoptimized + debuginfo] target(s) in 17.64s
[INFO] running `Command { std: "docker" "inspect" "b9c4e8843643e348ab92f0ace74247f38f0941a7f189e4d36c81e1d30617f9a8", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "b9c4e8843643e348ab92f0ace74247f38f0941a7f189e4d36c81e1d30617f9a8", kill_on_drop: false }`
[INFO] [stdout] b9c4e8843643e348ab92f0ace74247f38f0941a7f189e4d36c81e1d30617f9a8
