[INFO] cloning repository https://github.com/TuckerBMorgan/hermes
[INFO] running `Command { std: "git" "-c" "credential.helper=" "-c" "credential.helper=/workspace/cargo-home/bin/git-credential-null" "clone" "--bare" "https://github.com/TuckerBMorgan/hermes" "/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2FTuckerBMorgan%2Fhermes", kill_on_drop: false }`
[INFO] [stderr] Cloning into bare repository '/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2FTuckerBMorgan%2Fhermes'...
[INFO] running `Command { std: "git" "rev-parse" "HEAD", kill_on_drop: false }`
[INFO] [stdout] c08cc264a7d7f3a75373d1b5efd7bb440de78131
[INFO] building TuckerBMorgan/hermes against try#d79d87d4c0f7f0d0785ad6e493cf8cf718ed89a6 for pr-146237-1
[INFO] running `Command { std: "git" "clone" "/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2FTuckerBMorgan%2Fhermes" "/workspace/builds/worker-7-tc2/source", kill_on_drop: false }`
[INFO] [stderr] Cloning into '/workspace/builds/worker-7-tc2/source'...
[INFO] [stderr] done.
[INFO] started tweaking git repo https://github.com/TuckerBMorgan/hermes
[INFO] finished tweaking git repo https://github.com/TuckerBMorgan/hermes
[INFO] tweaked toml for git repo https://github.com/TuckerBMorgan/hermes written to /workspace/builds/worker-7-tc2/source/Cargo.toml
[INFO] validating manifest of git repo https://github.com/TuckerBMorgan/hermes on toolchain d79d87d4c0f7f0d0785ad6e493cf8cf718ed89a6
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+d79d87d4c0f7f0d0785ad6e493cf8cf718ed89a6" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }`
[INFO] crate git repo https://github.com/TuckerBMorgan/hermes already has a lockfile, it will not be regenerated
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+d79d87d4c0f7f0d0785ad6e493cf8cf718ed89a6" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-7-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-7-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:e90291280db7d1fac5b66fc6dad9f9662629e7365a55743daf9bdf73ebc4ea79" "/opt/rustwide/cargo-home/bin/cargo" "+d79d87d4c0f7f0d0785ad6e493cf8cf718ed89a6" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }`
[INFO] [stdout] 08dad828e1fff56ce468794ae0d51522052b24d94f91f69931a28b73986b8cf7
[INFO] running `Command { std: "docker" "start" "-a" "08dad828e1fff56ce468794ae0d51522052b24d94f91f69931a28b73986b8cf7", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "inspect" "08dad828e1fff56ce468794ae0d51522052b24d94f91f69931a28b73986b8cf7", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "08dad828e1fff56ce468794ae0d51522052b24d94f91f69931a28b73986b8cf7", kill_on_drop: false }`
[INFO] [stdout] 08dad828e1fff56ce468794ae0d51522052b24d94f91f69931a28b73986b8cf7
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-7-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-7-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:e90291280db7d1fac5b66fc6dad9f9662629e7365a55743daf9bdf73ebc4ea79" "/opt/rustwide/cargo-home/bin/cargo" "+d79d87d4c0f7f0d0785ad6e493cf8cf718ed89a6" "build" "--frozen" "--message-format=json", kill_on_drop: false }`
[INFO] [stdout] 3f4afb8717afa997a55d28f1fb6eaa47a722618a31939e13f2648cc76cdae95b
[INFO] running `Command { std: "docker" "start" "-a" "3f4afb8717afa997a55d28f1fb6eaa47a722618a31939e13f2648cc76cdae95b", kill_on_drop: false }`
[INFO] [stderr]    Compiling cc v1.2.7
[INFO] [stderr]    Compiling num-traits v0.2.19
[INFO] [stderr]    Compiling smallvec v1.13.2
[INFO] [stderr]    Compiling either v1.13.0
[INFO] [stderr]    Compiling openssl v0.10.68
[INFO] [stderr]    Compiling syn v2.0.95
[INFO] [stderr]    Compiling portable-atomic v1.10.0
[INFO] [stderr]    Compiling rustls v0.23.20
[INFO] [stderr]    Compiling percent-encoding v2.3.1
[INFO] [stderr]    Compiling unicode-width v0.2.0
[INFO] [stderr]    Compiling dirs-sys v0.4.1
[INFO] [stderr]    Compiling webpki-roots v0.26.7
[INFO] [stderr]    Compiling matrixmultiply v0.3.9
[INFO] [stderr]    Compiling aho-corasick v1.1.3
[INFO] [stderr]    Compiling number_prefix v0.4.0
[INFO] [stderr]    Compiling macro_rules_attribute-proc_macro v0.2.0
[INFO] [stderr]    Compiling unicode-segmentation v1.12.0
[INFO] [stderr]    Compiling unicode-normalization-alignments v0.1.12
[INFO] [stderr]    Compiling termcolor v1.1.3
[INFO] [stderr]    Compiling form_urlencoded v1.2.1
[INFO] [stderr]    Compiling rayon v1.10.0
[INFO] [stderr]    Compiling console v0.15.10
[INFO] [stderr]    Compiling dirs v5.0.1
[INFO] [stderr]    Compiling itertools v0.11.0
[INFO] [stderr]    Compiling itertools v0.12.1
[INFO] [stderr]    Compiling macro_rules_attribute v0.2.0
[INFO] [stderr]    Compiling num-integer v0.1.46
[INFO] [stderr]    Compiling num-complex v0.4.6
[INFO] [stderr]    Compiling chrono v0.4.39
[INFO] [stderr]    Compiling regex-automata v0.4.9
[INFO] [stderr]    Compiling indicatif v0.17.9
[INFO] [stderr]    Compiling openssl-sys v0.9.104
[INFO] [stderr]    Compiling ring v0.17.8
[INFO] [stderr]    Compiling onig_sys v69.8.1
[INFO] [stderr]    Compiling esaxx-rs v0.1.10
[INFO] [stderr]    Compiling rayon-cond v0.3.0
[INFO] [stderr]    Compiling native-tls v0.2.12
[INFO] [stderr]    Compiling ndarray v0.15.6
[INFO] [stderr]    Compiling simplelog v0.11.2
[INFO] [stderr]    Compiling synstructure v0.13.1
[INFO] [stderr]    Compiling darling_core v0.20.10
[INFO] [stderr]    Compiling regex v1.11.1
[INFO] [stderr]    Compiling zerovec-derive v0.10.3
[INFO] [stderr]    Compiling zerofrom-derive v0.1.5
[INFO] [stderr]    Compiling yoke-derive v0.7.5
[INFO] [stderr]    Compiling displaydoc v0.2.5
[INFO] [stderr]    Compiling icu_provider_macros v1.5.0
[INFO] [stderr]    Compiling serde_derive v1.0.217
[INFO] [stderr]    Compiling zerocopy-derive v0.7.35
[INFO] [stderr]    Compiling openssl-macros v0.1.1
[INFO] [stderr]    Compiling thiserror-impl v1.0.69
[INFO] [stderr]    Compiling monostate-impl v0.1.13
[INFO] [stderr]    Compiling zerocopy v0.7.35
[INFO] [stderr]    Compiling zerofrom v0.1.5
[INFO] [stderr]    Compiling yoke v0.7.5
[INFO] [stderr]    Compiling zerovec v0.10.4
[INFO] [stderr]    Compiling ppv-lite86 v0.2.20
[INFO] [stderr]    Compiling rustls-webpki v0.102.8
[INFO] [stderr]    Compiling darling_macro v0.20.10
[INFO] [stderr]    Compiling thiserror v1.0.69
[INFO] [stderr]    Compiling rand_chacha v0.3.1
[INFO] [stderr]    Compiling onig v6.4.0
[INFO] [stderr]    Compiling darling v0.20.10
[INFO] [stderr]    Compiling derive_builder_core v0.20.2
[INFO] [stderr]    Compiling tinystr v0.7.6
[INFO] [stderr]    Compiling icu_collections v1.5.0
[INFO] [stderr]    Compiling rand v0.8.5
[INFO] [stderr]    Compiling icu_locid v1.5.0
[INFO] [stderr]    Compiling icu_provider v1.5.0
[INFO] [stderr]    Compiling serde v1.0.217
[INFO] [stderr]    Compiling rand_distr v0.4.3
[INFO] [stderr]    Compiling icu_locid_transform v1.5.0
[INFO] [stderr]    Compiling derive_builder_macro v0.20.2
[INFO] [stderr]    Compiling icu_properties v1.5.1
[INFO] [stderr]    Compiling derive_builder v0.20.2
[INFO] [stderr]    Compiling icu_normalizer v1.5.0
[INFO] [stderr]    Compiling serde_json v1.0.135
[INFO] [stderr]    Compiling monostate v0.1.13
[INFO] [stderr]    Compiling spm_precompiled v0.1.4
[INFO] [stderr]    Compiling idna_adapter v1.2.0
[INFO] [stderr]    Compiling idna v1.0.3
[INFO] [stderr]    Compiling url v2.5.4
[INFO] [stderr]    Compiling ureq v2.12.1
[INFO] [stderr]    Compiling hf-hub v0.3.2
[INFO] [stderr]    Compiling tokenizers v0.20.4
[INFO] [stderr]    Compiling poro v0.1.2 (/opt/rustwide/workdir)
[INFO] [stdout] warning: unused import: `ndarray::prelude::*`
[INFO] [stdout]  --> src/central/add_op.rs:7:9
[INFO] [stdout]   |
[INFO] [stdout] 7 | pub use ndarray::prelude::*;
[INFO] [stdout]   |         ^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout]   = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `Result`
[INFO] [stdout]  --> src/central/tensor.rs:7:21
[INFO] [stdout]   |
[INFO] [stdout] 7 | use std::io::{Read, Result};
[INFO] [stdout]   |                     ^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unnecessary trailing semicolon
[INFO] [stdout]    --> src/central/tensor.rs:224:92
[INFO] [stdout]     |
[INFO] [stdout] 224 |             .map(|chunk| i32::from_le_bytes(chunk.try_into().unwrap()) as usize).collect();;
[INFO] [stdout]     |                                                                                            ^ help: remove this semicolon
[INFO] [stdout]     |
[INFO] [stdout]     = note: `#[warn(redundant_semicolons)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused imports: `LinearLayerConfig` and `LinearLayer`
[INFO] [stdout]   --> src/nn/layers/linear.rs:55:37
[INFO] [stdout]    |
[INFO] [stdout] 55 |     use crate::nn::layers::linear::{LinearLayer, LinearLayerConfig};
[INFO] [stdout]    |                                     ^^^^^^^^^^^  ^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::nn::layers::module::Module`
[INFO] [stdout]   --> src/nn/layers/linear.rs:56:9
[INFO] [stdout]    |
[INFO] [stdout] 56 |     use crate::nn::layers::module::Module;
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::central::Tensor`
[INFO] [stdout]   --> src/nn/layers/linear.rs:57:9
[INFO] [stdout]    |
[INFO] [stdout] 57 |     use crate::central::Tensor;
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::Shape`
[INFO] [stdout]   --> src/nn/layers/linear.rs:58:9
[INFO] [stdout]    |
[INFO] [stdout] 58 |     use crate::Shape;
[INFO] [stdout]    |         ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::nn::layers::linear`
[INFO] [stdout]   --> src/nn/layers/mlp.rs:80:9
[INFO] [stdout]    |
[INFO] [stdout] 80 |     use crate::nn::layers::linear;
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::nn::layers::mlp::MLP`
[INFO] [stdout]   --> src/nn/layers/mlp.rs:81:9
[INFO] [stdout]    |
[INFO] [stdout] 81 |     use crate::nn::layers::mlp::MLP;
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::nn::layers::module::Module`
[INFO] [stdout]   --> src/nn/layers/mlp.rs:82:9
[INFO] [stdout]    |
[INFO] [stdout] 82 |     use crate::nn::layers::module::Module;
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::central::Tensor`
[INFO] [stdout]   --> src/nn/layers/mlp.rs:83:9
[INFO] [stdout]    |
[INFO] [stdout] 83 |     use crate::central::Tensor;
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::Shape`
[INFO] [stdout]   --> src/nn/layers/mlp.rs:84:9
[INFO] [stdout]    |
[INFO] [stdout] 84 |     use crate::Shape;
[INFO] [stdout]    |         ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `core::panic`
[INFO] [stdout]  --> src/nn/layers/layer_norm.rs:1:5
[INFO] [stdout]   |
[INFO] [stdout] 1 | use core::panic;
[INFO] [stdout]   |     ^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unnecessary trailing semicolon
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:56:47
[INFO] [stdout]    |
[INFO] [stdout] 56 |         let var: Tensor = x.var(mean_indices);;//(input_minus_mean - input_minus_mean).pow(2.0).mean(mean_indices.clone());
[INFO] [stdout]    |                                               ^ help: remove this semicolon
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `core::panic`
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:77:9
[INFO] [stdout]    |
[INFO] [stdout] 77 |     use core::panic;
[INFO] [stdout]    |         ^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `ndarray::prelude::*`
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:79:9
[INFO] [stdout]    |
[INFO] [stdout] 79 |     use ndarray::prelude::*;
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `ndarray::iter::Axes`
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:87:9
[INFO] [stdout]    |
[INFO] [stdout] 87 |     use ndarray::iter::Axes;
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `ndarray::Axis`
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:88:9
[INFO] [stdout]    |
[INFO] [stdout] 88 |     use ndarray::Axis;
[INFO] [stdout]    |         ^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::central::Tensor`
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:90:9
[INFO] [stdout]    |
[INFO] [stdout] 90 |     use crate::central::Tensor;
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::nn::layers::layer_norm::LayerNorm`
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:91:9
[INFO] [stdout]    |
[INFO] [stdout] 91 |     use crate::nn::layers::layer_norm::LayerNorm;
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::nn::layers::module::Module`
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:92:9
[INFO] [stdout]    |
[INFO] [stdout] 92 |     use crate::nn::layers::module::Module;
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `simplelog::*`
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:93:9
[INFO] [stdout]    |
[INFO] [stdout] 93 |     use simplelog::*;
[INFO] [stdout]    |         ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `log::info`
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:94:9
[INFO] [stdout]    |
[INFO] [stdout] 94 |     use log::info;
[INFO] [stdout]    |         ^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `std::path::Path`
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:95:9
[INFO] [stdout]    |
[INFO] [stdout] 95 |     use std::path::Path;    
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `std::fs::File`
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:96:9
[INFO] [stdout]    |
[INFO] [stdout] 96 |     use std::fs::File;
[INFO] [stdout]    |         ^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `std::vec`
[INFO] [stdout]    --> src/nn/layers/casual_self_attention.rs:203:9
[INFO] [stdout]     |
[INFO] [stdout] 203 |     use std::vec;
[INFO] [stdout]     |         ^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::nn::layers::linear::LinearLayerConfig`
[INFO] [stdout]    --> src/nn/layers/casual_self_attention.rs:204:9
[INFO] [stdout]     |
[INFO] [stdout] 204 |     use crate::nn::layers::linear::LinearLayerConfig;
[INFO] [stdout]     |         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::nn::layers::module::Module`
[INFO] [stdout]    --> src/nn/layers/casual_self_attention.rs:205:9
[INFO] [stdout]     |
[INFO] [stdout] 205 |     use crate::nn::layers::module::Module;
[INFO] [stdout]     |         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::central::Tensor`
[INFO] [stdout]    --> src/nn/layers/casual_self_attention.rs:206:9
[INFO] [stdout]     |
[INFO] [stdout] 206 |     use crate::central::Tensor;
[INFO] [stdout]     |         ^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::Shape`
[INFO] [stdout]    --> src/nn/layers/casual_self_attention.rs:207:9
[INFO] [stdout]     |
[INFO] [stdout] 207 |     use crate::Shape;
[INFO] [stdout]     |         ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `std::hash::Hash`
[INFO] [stdout]  --> src/optimizers/adam.rs:2:5
[INFO] [stdout]   |
[INFO] [stdout] 2 | use std::hash::Hash;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `std::hash::Hash`
[INFO] [stdout]  --> src/optimizers/adamw.rs:2:5
[INFO] [stdout]   |
[INFO] [stdout] 2 | use std::hash::Hash;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unreachable statement
[INFO] [stdout]   --> src/nn/layers/embedding.rs:31:9
[INFO] [stdout]    |
[INFO] [stdout] 29 |         return self.tensor.embbeding(input, self.model_dimension);
[INFO] [stdout]    |         --------------------------------------------------------- any code following this expression is unreachable
[INFO] [stdout] 30 |         
[INFO] [stdout] 31 |         let mut test_index_tensor = Tensor::zeroes(Shape::new(vec![input.shape.indices[0], input.shape.indices[1], self.model_dimension]));
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ unreachable statement
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(unreachable_code)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `source_grad`
[INFO] [stdout]    --> src/central/equation.rs:805:21
[INFO] [stdout]     |
[INFO] [stdout] 805 |                 let source_grad = self.get_tensor_grad(origin_tensor).into_raw_vec();
[INFO] [stdout]     |                     ^^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_source_grad`
[INFO] [stdout]     |
[INFO] [stdout]     = note: `#[warn(unused_variables)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: variable `M` should have a snake case name
[INFO] [stdout]    --> src/central/equation.rs:431:9
[INFO] [stdout]     |
[INFO] [stdout] 431 |     let M = a_shape[a_shape.len() - 2];
[INFO] [stdout]     |         ^ help: convert the identifier to snake case: `m`
[INFO] [stdout]     |
[INFO] [stdout]     = note: `#[warn(non_snake_case)]` (part of `#[warn(nonstandard_style)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: variable `K_a` should have a snake case name
[INFO] [stdout]    --> src/central/equation.rs:432:9
[INFO] [stdout]     |
[INFO] [stdout] 432 |     let K_a = a_shape[a_shape.len() - 1];
[INFO] [stdout]     |         ^^^ help: convert the identifier to snake case (notice the capitalization): `k_a`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: variable `K_b` should have a snake case name
[INFO] [stdout]    --> src/central/equation.rs:433:9
[INFO] [stdout]     |
[INFO] [stdout] 433 |     let K_b = b_shape[b_shape.len() - 2];
[INFO] [stdout]     |         ^^^ help: convert the identifier to snake case (notice the capitalization): `k_b`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: variable `N` should have a snake case name
[INFO] [stdout]    --> src/central/equation.rs:434:9
[INFO] [stdout]     |
[INFO] [stdout] 434 |     let N = b_shape[b_shape.len() - 1];
[INFO] [stdout]     |         ^ help: convert the identifier to snake case: `n`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: method `add_time` is never used
[INFO] [stdout]    --> src/central/equation.rs:201:8
[INFO] [stdout]     |
[INFO] [stdout]  46 | impl Equation {
[INFO] [stdout]     | ------------- method in this implementation
[INFO] [stdout] ...
[INFO] [stdout] 201 |     fn add_time(&mut self, operation: &str, now: Instant) {
[INFO] [stdout]     |        ^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout]     = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `transpose_inside` is never used
[INFO] [stdout]    --> src/central/tensor.rs:632:8
[INFO] [stdout]     |
[INFO] [stdout]  45 | impl Tensor {
[INFO] [stdout]     | ----------- associated function in this implementation
[INFO] [stdout] ...
[INFO] [stdout] 632 |     fn transpose_inside(matrix: Vec<f32>, rows: usize, cols: usize) -> Vec<f32> {
[INFO] [stdout]     |        ^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: struct `MLPConfig` is never constructed
[INFO] [stdout]   --> src/nn/layers/mlp.rs:23:8
[INFO] [stdout]    |
[INFO] [stdout] 23 | struct MLPConfig {
[INFO] [stdout]    |        ^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `new` is never used
[INFO] [stdout]   --> src/nn/layers/mlp.rs:34:8
[INFO] [stdout]    |
[INFO] [stdout] 33 | impl MLP {
[INFO] [stdout]    | -------- associated function in this implementation
[INFO] [stdout] 34 |     fn new(config: MLPConfig) -> Self {
[INFO] [stdout]    |        ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: field `length_of_normalized_shape` is never read
[INFO] [stdout]  --> src/nn/layers/layer_norm.rs:9:5
[INFO] [stdout]   |
[INFO] [stdout] 6 | pub struct LayerNorm {
[INFO] [stdout]   |            --------- field in this struct
[INFO] [stdout] ...
[INFO] [stdout] 9 |     length_of_normalized_shape: usize,
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `write_f32_vector_to_file` is never used
[INFO] [stdout]  --> src/nn/layers/casual_self_attention.rs:7:4
[INFO] [stdout]   |
[INFO] [stdout] 7 | fn write_f32_vector_to_file(path: &str, data: &[f32]) -> std::io::Result<()> {
[INFO] [stdout]   |    ^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `write_string_vector_to_file` is never used
[INFO] [stdout]   --> src/nn/layers/casual_self_attention.rs:29:4
[INFO] [stdout]    |
[INFO] [stdout] 29 | fn write_string_vector_to_file(path: &str, data: &str) -> std::io::Result<()> {
[INFO] [stdout]    |    ^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: field `vocab_size` is never read
[INFO] [stdout]  --> src/nn/layers/embedding.rs:7:5
[INFO] [stdout]   |
[INFO] [stdout] 5 | pub struct Embedding {
[INFO] [stdout]   |            --------- field in this struct
[INFO] [stdout] 6 |     pub tensor: Tensor,
[INFO] [stdout] 7 |     vocab_size: usize,
[INFO] [stdout]   |     ^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stderr]     Finished `dev` profile [unoptimized + debuginfo] target(s) in 1m 15s
[INFO] running `Command { std: "docker" "inspect" "3f4afb8717afa997a55d28f1fb6eaa47a722618a31939e13f2648cc76cdae95b", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "3f4afb8717afa997a55d28f1fb6eaa47a722618a31939e13f2648cc76cdae95b", kill_on_drop: false }`
[INFO] [stdout] 3f4afb8717afa997a55d28f1fb6eaa47a722618a31939e13f2648cc76cdae95b
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-7-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-7-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:e90291280db7d1fac5b66fc6dad9f9662629e7365a55743daf9bdf73ebc4ea79" "/opt/rustwide/cargo-home/bin/cargo" "+d79d87d4c0f7f0d0785ad6e493cf8cf718ed89a6" "test" "--frozen" "--no-run" "--message-format=json", kill_on_drop: false }`
[INFO] [stdout] 7edaba70dea76f816c8a4506f2d5e176c9b84f2c72c1fdba07bbe12cb7796183
[INFO] running `Command { std: "docker" "start" "-a" "7edaba70dea76f816c8a4506f2d5e176c9b84f2c72c1fdba07bbe12cb7796183", kill_on_drop: false }`
[INFO] [stdout] warning: unused import: `ndarray::prelude::*`
[INFO] [stdout]  --> src/central/add_op.rs:7:9
[INFO] [stdout]   |
[INFO] [stdout] 7 | pub use ndarray::prelude::*;
[INFO] [stdout]   |         ^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout]   = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `Result`
[INFO] [stdout]  --> src/central/tensor.rs:7:21
[INFO] [stdout]   |
[INFO] [stdout] 7 | use std::io::{Read, Result};
[INFO] [stdout]   |                     ^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unnecessary trailing semicolon
[INFO] [stdout]    --> src/central/tensor.rs:224:92
[INFO] [stdout]     |
[INFO] [stdout] 224 |             .map(|chunk| i32::from_le_bytes(chunk.try_into().unwrap()) as usize).collect();;
[INFO] [stdout]     |                                                                                            ^ help: remove this semicolon
[INFO] [stdout]     |
[INFO] [stdout]     = note: `#[warn(redundant_semicolons)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused imports: `LinearLayerConfig` and `LinearLayer`
[INFO] [stdout]   --> src/nn/layers/linear.rs:55:37
[INFO] [stdout]    |
[INFO] [stdout] 55 |     use crate::nn::layers::linear::{LinearLayer, LinearLayerConfig};
[INFO] [stdout]    |                                     ^^^^^^^^^^^  ^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::nn::layers::module::Module`
[INFO] [stdout]   --> src/nn/layers/linear.rs:56:9
[INFO] [stdout]    |
[INFO] [stdout] 56 |     use crate::nn::layers::module::Module;
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::central::Tensor`
[INFO] [stdout]   --> src/nn/layers/linear.rs:57:9
[INFO] [stdout]    |
[INFO] [stdout] 57 |     use crate::central::Tensor;
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::Shape`
[INFO] [stdout]   --> src/nn/layers/linear.rs:58:9
[INFO] [stdout]    |
[INFO] [stdout] 58 |     use crate::Shape;
[INFO] [stdout]    |         ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::nn::layers::linear`
[INFO] [stdout]   --> src/nn/layers/mlp.rs:80:9
[INFO] [stdout]    |
[INFO] [stdout] 80 |     use crate::nn::layers::linear;
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::nn::layers::mlp::MLP`
[INFO] [stdout]   --> src/nn/layers/mlp.rs:81:9
[INFO] [stdout]    |
[INFO] [stdout] 81 |     use crate::nn::layers::mlp::MLP;
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::nn::layers::module::Module`
[INFO] [stdout]   --> src/nn/layers/mlp.rs:82:9
[INFO] [stdout]    |
[INFO] [stdout] 82 |     use crate::nn::layers::module::Module;
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::central::Tensor`
[INFO] [stdout]   --> src/nn/layers/mlp.rs:83:9
[INFO] [stdout]    |
[INFO] [stdout] 83 |     use crate::central::Tensor;
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::Shape`
[INFO] [stdout]   --> src/nn/layers/mlp.rs:84:9
[INFO] [stdout]    |
[INFO] [stdout] 84 |     use crate::Shape;
[INFO] [stdout]    |         ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `core::panic`
[INFO] [stdout]  --> src/nn/layers/layer_norm.rs:1:5
[INFO] [stdout]   |
[INFO] [stdout] 1 | use core::panic;
[INFO] [stdout]   |     ^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unnecessary trailing semicolon
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:56:47
[INFO] [stdout]    |
[INFO] [stdout] 56 |         let var: Tensor = x.var(mean_indices);;//(input_minus_mean - input_minus_mean).pow(2.0).mean(mean_indices.clone());
[INFO] [stdout]    |                                               ^ help: remove this semicolon
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `core::panic`
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:77:9
[INFO] [stdout]    |
[INFO] [stdout] 77 |     use core::panic;
[INFO] [stdout]    |         ^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `ndarray::prelude::*`
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:79:9
[INFO] [stdout]    |
[INFO] [stdout] 79 |     use ndarray::prelude::*;
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `ndarray::iter::Axes`
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:87:9
[INFO] [stdout]    |
[INFO] [stdout] 87 |     use ndarray::iter::Axes;
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `ndarray::Axis`
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:88:9
[INFO] [stdout]    |
[INFO] [stdout] 88 |     use ndarray::Axis;
[INFO] [stdout]    |         ^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::central::Tensor`
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:90:9
[INFO] [stdout]    |
[INFO] [stdout] 90 |     use crate::central::Tensor;
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::nn::layers::layer_norm::LayerNorm`
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:91:9
[INFO] [stdout]    |
[INFO] [stdout] 91 |     use crate::nn::layers::layer_norm::LayerNorm;
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::nn::layers::module::Module`
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:92:9
[INFO] [stdout]    |
[INFO] [stdout] 92 |     use crate::nn::layers::module::Module;
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `simplelog::*`
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:93:9
[INFO] [stdout]    |
[INFO] [stdout] 93 |     use simplelog::*;
[INFO] [stdout]    |         ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `log::info`
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:94:9
[INFO] [stdout]    |
[INFO] [stdout] 94 |     use log::info;
[INFO] [stdout]    |         ^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `std::path::Path`
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:95:9
[INFO] [stdout]    |
[INFO] [stdout] 95 |     use std::path::Path;    
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `std::fs::File`
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:96:9
[INFO] [stdout]    |
[INFO] [stdout] 96 |     use std::fs::File;
[INFO] [stdout]    |         ^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `std::vec`
[INFO] [stdout]    --> src/nn/layers/casual_self_attention.rs:203:9
[INFO] [stdout]     |
[INFO] [stdout] 203 |     use std::vec;
[INFO] [stdout]     |         ^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::nn::layers::linear::LinearLayerConfig`
[INFO] [stdout]    --> src/nn/layers/casual_self_attention.rs:204:9
[INFO] [stdout]     |
[INFO] [stdout] 204 |     use crate::nn::layers::linear::LinearLayerConfig;
[INFO] [stdout]     |         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::nn::layers::module::Module`
[INFO] [stdout]    --> src/nn/layers/casual_self_attention.rs:205:9
[INFO] [stdout]     |
[INFO] [stdout] 205 |     use crate::nn::layers::module::Module;
[INFO] [stdout]     |         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::central::Tensor`
[INFO] [stdout]    --> src/nn/layers/casual_self_attention.rs:206:9
[INFO] [stdout]     |
[INFO] [stdout] 206 |     use crate::central::Tensor;
[INFO] [stdout]     |         ^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::Shape`
[INFO] [stdout]    --> src/nn/layers/casual_self_attention.rs:207:9
[INFO] [stdout]     |
[INFO] [stdout] 207 |     use crate::Shape;
[INFO] [stdout]     |         ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `std::hash::Hash`
[INFO] [stdout]  --> src/optimizers/adam.rs:2:5
[INFO] [stdout]   |
[INFO] [stdout] 2 | use std::hash::Hash;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `std::hash::Hash`
[INFO] [stdout]  --> src/optimizers/adamw.rs:2:5
[INFO] [stdout]   |
[INFO] [stdout] 2 | use std::hash::Hash;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unreachable statement
[INFO] [stdout]   --> src/nn/layers/embedding.rs:31:9
[INFO] [stdout]    |
[INFO] [stdout] 29 |         return self.tensor.embbeding(input, self.model_dimension);
[INFO] [stdout]    |         --------------------------------------------------------- any code following this expression is unreachable
[INFO] [stdout] 30 |         
[INFO] [stdout] 31 |         let mut test_index_tensor = Tensor::zeroes(Shape::new(vec![input.shape.indices[0], input.shape.indices[1], self.model_dimension]));
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ unreachable statement
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(unreachable_code)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `source_grad`
[INFO] [stdout]    --> src/central/equation.rs:805:21
[INFO] [stdout]     |
[INFO] [stdout] 805 |                 let source_grad = self.get_tensor_grad(origin_tensor).into_raw_vec();
[INFO] [stdout]     |                     ^^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_source_grad`
[INFO] [stdout]     |
[INFO] [stdout]     = note: `#[warn(unused_variables)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: variable `M` should have a snake case name
[INFO] [stdout]    --> src/central/equation.rs:431:9
[INFO] [stdout]     |
[INFO] [stdout] 431 |     let M = a_shape[a_shape.len() - 2];
[INFO] [stdout]     |         ^ help: convert the identifier to snake case: `m`
[INFO] [stdout]     |
[INFO] [stdout]     = note: `#[warn(non_snake_case)]` (part of `#[warn(nonstandard_style)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stderr]    Compiling poro v0.1.2 (/opt/rustwide/workdir)
[INFO] [stdout] warning: variable `K_a` should have a snake case name
[INFO] [stdout]    --> src/central/equation.rs:432:9
[INFO] [stdout]     |
[INFO] [stdout] 432 |     let K_a = a_shape[a_shape.len() - 1];
[INFO] [stdout]     |         ^^^ help: convert the identifier to snake case (notice the capitalization): `k_a`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: variable `K_b` should have a snake case name
[INFO] [stdout]    --> src/central/equation.rs:433:9
[INFO] [stdout]     |
[INFO] [stdout] 433 |     let K_b = b_shape[b_shape.len() - 2];
[INFO] [stdout]     |         ^^^ help: convert the identifier to snake case (notice the capitalization): `k_b`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: variable `N` should have a snake case name
[INFO] [stdout]    --> src/central/equation.rs:434:9
[INFO] [stdout]     |
[INFO] [stdout] 434 |     let N = b_shape[b_shape.len() - 1];
[INFO] [stdout]     |         ^ help: convert the identifier to snake case: `n`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: method `add_time` is never used
[INFO] [stdout]    --> src/central/equation.rs:201:8
[INFO] [stdout]     |
[INFO] [stdout]  46 | impl Equation {
[INFO] [stdout]     | ------------- method in this implementation
[INFO] [stdout] ...
[INFO] [stdout] 201 |     fn add_time(&mut self, operation: &str, now: Instant) {
[INFO] [stdout]     |        ^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout]     = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `transpose_inside` is never used
[INFO] [stdout]    --> src/central/tensor.rs:632:8
[INFO] [stdout]     |
[INFO] [stdout]  45 | impl Tensor {
[INFO] [stdout]     | ----------- associated function in this implementation
[INFO] [stdout] ...
[INFO] [stdout] 632 |     fn transpose_inside(matrix: Vec<f32>, rows: usize, cols: usize) -> Vec<f32> {
[INFO] [stdout]     |        ^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: struct `MLPConfig` is never constructed
[INFO] [stdout]   --> src/nn/layers/mlp.rs:23:8
[INFO] [stdout]    |
[INFO] [stdout] 23 | struct MLPConfig {
[INFO] [stdout]    |        ^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `new` is never used
[INFO] [stdout]   --> src/nn/layers/mlp.rs:34:8
[INFO] [stdout]    |
[INFO] [stdout] 33 | impl MLP {
[INFO] [stdout]    | -------- associated function in this implementation
[INFO] [stdout] 34 |     fn new(config: MLPConfig) -> Self {
[INFO] [stdout]    |        ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: field `length_of_normalized_shape` is never read
[INFO] [stdout]  --> src/nn/layers/layer_norm.rs:9:5
[INFO] [stdout]   |
[INFO] [stdout] 6 | pub struct LayerNorm {
[INFO] [stdout]   |            --------- field in this struct
[INFO] [stdout] ...
[INFO] [stdout] 9 |     length_of_normalized_shape: usize,
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `write_f32_vector_to_file` is never used
[INFO] [stdout]  --> src/nn/layers/casual_self_attention.rs:7:4
[INFO] [stdout]   |
[INFO] [stdout] 7 | fn write_f32_vector_to_file(path: &str, data: &[f32]) -> std::io::Result<()> {
[INFO] [stdout]   |    ^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `write_string_vector_to_file` is never used
[INFO] [stdout]   --> src/nn/layers/casual_self_attention.rs:29:4
[INFO] [stdout]    |
[INFO] [stdout] 29 | fn write_string_vector_to_file(path: &str, data: &str) -> std::io::Result<()> {
[INFO] [stdout]    |    ^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: field `vocab_size` is never read
[INFO] [stdout]  --> src/nn/layers/embedding.rs:7:5
[INFO] [stdout]   |
[INFO] [stdout] 5 | pub struct Embedding {
[INFO] [stdout]   |            --------- field in this struct
[INFO] [stdout] 6 |     pub tensor: Tensor,
[INFO] [stdout] 7 |     vocab_size: usize,
[INFO] [stdout]   |     ^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `from_tensor` is never used
[INFO] [stdout]   --> examples/wavenet.rs:20:12
[INFO] [stdout]    |
[INFO] [stdout] 13 | impl EmbeddingLayer {
[INFO] [stdout]    | ------------------- associated function in this implementation
[INFO] [stdout] ...
[INFO] [stdout] 20 |     pub fn from_tensor(weights: Tensor) -> Self {
[INFO] [stdout]    |            ^^^^^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused imports: `AdamBuilder`, `Adam`, and `LinearLayerConfig`
[INFO] [stdout]  --> examples/linear_and_adam_optimizer.rs:1:44
[INFO] [stdout]   |
[INFO] [stdout] 1 | use poro::{get_equation, nn::{LinearLayer, LinearLayerConfig}, Adam, AdamBuilder, AdamW, AdamWBuilder, Optimizer};
[INFO] [stdout]   |                                            ^^^^^^^^^^^^^^^^^   ^^^^  ^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout]   = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `Result`
[INFO] [stdout]  --> src/central/tensor.rs:7:21
[INFO] [stdout]   |
[INFO] [stdout] 7 | use std::io::{Read, Result};
[INFO] [stdout]   |                     ^^^^^^
[INFO] [stdout]   |
[INFO] [stdout]   = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unnecessary trailing semicolon
[INFO] [stdout]    --> src/central/tensor.rs:224:92
[INFO] [stdout]     |
[INFO] [stdout] 224 |             .map(|chunk| i32::from_le_bytes(chunk.try_into().unwrap()) as usize).collect();;
[INFO] [stdout]     |                                                                                            ^ help: remove this semicolon
[INFO] [stdout]     |
[INFO] [stdout]     = note: `#[warn(redundant_semicolons)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::Shape`
[INFO] [stdout]   --> src/nn/layers/linear.rs:58:9
[INFO] [stdout]    |
[INFO] [stdout] 58 |     use crate::Shape;
[INFO] [stdout]    |         ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::nn::layers::linear`
[INFO] [stdout]   --> src/nn/layers/mlp.rs:80:9
[INFO] [stdout]    |
[INFO] [stdout] 80 |     use crate::nn::layers::linear;
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::Shape`
[INFO] [stdout]   --> src/nn/layers/mlp.rs:84:9
[INFO] [stdout]    |
[INFO] [stdout] 84 |     use crate::Shape;
[INFO] [stdout]    |         ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `core::panic`
[INFO] [stdout]  --> src/nn/layers/layer_norm.rs:1:5
[INFO] [stdout]   |
[INFO] [stdout] 1 | use core::panic;
[INFO] [stdout]   |     ^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unnecessary trailing semicolon
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:56:47
[INFO] [stdout]    |
[INFO] [stdout] 56 |         let var: Tensor = x.var(mean_indices);;//(input_minus_mean - input_minus_mean).pow(2.0).mean(mean_indices.clone());
[INFO] [stdout]    |                                               ^ help: remove this semicolon
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `ndarray::iter::Axes`
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:87:9
[INFO] [stdout]    |
[INFO] [stdout] 87 |     use ndarray::iter::Axes;
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `ndarray::Axis`
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:88:9
[INFO] [stdout]    |
[INFO] [stdout] 88 |     use ndarray::Axis;
[INFO] [stdout]    |         ^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `log::info`
[INFO] [stdout]   --> src/nn/layers/layer_norm.rs:94:9
[INFO] [stdout]    |
[INFO] [stdout] 94 |     use log::info;
[INFO] [stdout]    |         ^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::nn::layers::linear::LinearLayerConfig`
[INFO] [stdout]    --> src/nn/layers/casual_self_attention.rs:204:9
[INFO] [stdout]     |
[INFO] [stdout] 204 |     use crate::nn::layers::linear::LinearLayerConfig;
[INFO] [stdout]     |         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::central::Tensor`
[INFO] [stdout]    --> src/nn/layers/casual_self_attention.rs:206:9
[INFO] [stdout]     |
[INFO] [stdout] 206 |     use crate::central::Tensor;
[INFO] [stdout]     |         ^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::Shape`
[INFO] [stdout]    --> src/nn/layers/casual_self_attention.rs:207:9
[INFO] [stdout]     |
[INFO] [stdout] 207 |     use crate::Shape;
[INFO] [stdout]     |         ^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `std::hash::Hash`
[INFO] [stdout]  --> src/optimizers/adam.rs:2:5
[INFO] [stdout]   |
[INFO] [stdout] 2 | use std::hash::Hash;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `std::hash::Hash`
[INFO] [stdout]  --> src/optimizers/adamw.rs:2:5
[INFO] [stdout]   |
[INFO] [stdout] 2 | use std::hash::Hash;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `CasualSelfAttentionConfig`
[INFO] [stdout]  --> examples/transfomer_test.rs:4:107
[INFO] [stdout]   |
[INFO] [stdout] 4 | use poro::nn::{AttentionHead, Embedding, LinearLayer, Model, Module, MLP, LayerNorm, CasualSelfAttention, CasualSelfAttentionConfig};
[INFO] [stdout]   |                                                                                                           ^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout]   = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `poro::optimizers::adam::Adam`
[INFO] [stdout]  --> examples/transfomer_test.rs:6:5
[INFO] [stdout]   |
[INFO] [stdout] 6 | use poro::optimizers::adam::Adam;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `poro::nn::layers::NewGLU`
[INFO] [stdout]  --> examples/transfomer_test.rs:7:5
[INFO] [stdout]   |
[INFO] [stdout] 7 | use poro::nn::layers::NewGLU;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `Seek`
[INFO] [stdout]  --> examples/transfomer_test.rs:8:21
[INFO] [stdout]   |
[INFO] [stdout] 8 | use std::io::{Read, Seek};
[INFO] [stdout]   |                     ^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused imports: `Result` and `Tokenizer`
[INFO] [stdout]   --> examples/transfomer_test.rs:16:29
[INFO] [stdout]    |
[INFO] [stdout] 16 | use tokenizers::tokenizer::{Result, Tokenizer};
[INFO] [stdout]    |                             ^^^^^^  ^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `update_parameters`
[INFO] [stdout]    --> examples/transfomer_test.rs:398:12
[INFO] [stdout]     |
[INFO] [stdout] 398 | use poro::{update_parameters, Shape};
[INFO] [stdout]     |            ^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `rand::Rng`
[INFO] [stdout]    --> examples/transfomer_test.rs:399:5
[INFO] [stdout]     |
[INFO] [stdout] 399 | use rand::Rng;
[INFO] [stdout]     |     ^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unreachable statement
[INFO] [stdout]    --> examples/transfomer_test.rs:532:5
[INFO] [stdout]     |
[INFO] [stdout] 530 |     panic!("");
[INFO] [stdout]     |     ---------- any code following this expression is unreachable
[INFO] [stdout] 531 |
[INFO] [stdout] 532 |     let test_output = test_ouput.reshape(vec![batch_size * seq_length, 50257].into());
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ unreachable statement
[INFO] [stdout]     |
[INFO] [stdout]     = note: `#[warn(unreachable_code)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `as_vec`
[INFO] [stdout]    --> examples/transfomer_test.rs:417:14
[INFO] [stdout]     |
[INFO] [stdout] 417 |         let  as_vec = read.read_to_end(&mut vec_buffer).unwrap();
[INFO] [stdout]     |              ^^^^^^ help: if this is intentional, prefix it with an underscore: `_as_vec`
[INFO] [stdout]     |
[INFO] [stdout]     = note: `#[warn(unused_variables)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `vocab_size`
[INFO] [stdout]    --> examples/transfomer_test.rs:508:9
[INFO] [stdout]     |
[INFO] [stdout] 508 |     let vocab_size = 50304;
[INFO] [stdout]     |         ^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_vocab_size`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `one_hot_encoded_trues`
[INFO] [stdout]    --> examples/transfomer_test.rs:525:9
[INFO] [stdout]     |
[INFO] [stdout] 525 |     let one_hot_encoded_trues = one_hot_encoded_trues.reshape(vec![batch_size *  seq_length, 50257].into());
[INFO] [stdout]     |         ^^^^^^^^^^^^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_one_hot_encoded_trues`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `test_ouput`
[INFO] [stdout]    --> examples/transfomer_test.rs:528:9
[INFO] [stdout]     |
[INFO] [stdout] 528 |     let test_ouput = gpt.forward(&x);
[INFO] [stdout]     |         ^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_test_ouput`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `input`
[INFO] [stdout]    --> examples/transfomer_test.rs:565:9
[INFO] [stdout]     |
[INFO] [stdout] 565 |     let input = Tensor::randn(vec![1, 1, 1].into());
[INFO] [stdout]     |         ^^^^^ help: if this is intentional, prefix it with an underscore: `_input`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `output`
[INFO] [stdout]    --> examples/transfomer_test.rs:566:9
[INFO] [stdout]     |
[INFO] [stdout] 566 |     let output = gpt.forward(&test_input);
[INFO] [stdout]     |         ^^^^^^ help: if this is intentional, prefix it with an underscore: `_output`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `gpt_config`
[INFO] [stdout]    --> examples/transfomer_test.rs:568:9
[INFO] [stdout]     |
[INFO] [stdout] 568 |     let gpt_config = GPTConfig::from_bytestream(&mut std::fs::File::open("gpt2.bin").unwrap());
[INFO] [stdout]     |         ^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_gpt_config`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `position_encoder`
[INFO] [stdout]    --> examples/transfomer_test.rs:572:13
[INFO] [stdout]     |
[INFO] [stdout] 572 |     let mut position_encoder = PositionalEncoding::new(10, 10);    
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_position_encoder`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `embeddings`
[INFO] [stdout]    --> examples/transfomer_test.rs:587:13
[INFO] [stdout]     |
[INFO] [stdout] 587 |     let mut embeddings = Embedding::from_tensor(embeddings_weight);
[INFO] [stdout]     |             ^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_embeddings`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `write_f32_vector_to_file` is never used
[INFO] [stdout]   --> examples/transfomer_test.rs:17:4
[INFO] [stdout]    |
[INFO] [stdout] 17 | fn write_f32_vector_to_file(path: &str, data: &[f32]) -> std::io::Result<()> {
[INFO] [stdout]    |    ^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `write_string_vector_to_file` is never used
[INFO] [stdout]   --> examples/transfomer_test.rs:39:4
[INFO] [stdout]    |
[INFO] [stdout] 39 | fn write_string_vector_to_file(path: &str, data: &str) -> std::io::Result<()> {
[INFO] [stdout]    |    ^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: fields `block_size`, `vocab_size`, `number_of_heads`, and `embedding_dim` are never read
[INFO] [stdout]    --> examples/transfomer_test.rs:130:5
[INFO] [stdout]     |
[INFO] [stdout] 129 | struct GPTConfig {
[INFO] [stdout]     |        --------- fields in this struct
[INFO] [stdout] 130 |     block_size: usize,
[INFO] [stdout]     |     ^^^^^^^^^^
[INFO] [stdout] 131 |     vocab_size: usize,
[INFO] [stdout]     |     ^^^^^^^^^^
[INFO] [stdout] 132 |     number_of_layers: usize,
[INFO] [stdout] 133 |     number_of_heads: usize,
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^
[INFO] [stdout] 134 |     embedding_dim: usize  
[INFO] [stdout]     |     ^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `get_learning_rate` is never used
[INFO] [stdout]    --> examples/transfomer_test.rs:477:4
[INFO] [stdout]     |
[INFO] [stdout] 477 | fn get_learning_rate(iteration: usize) -> f32 {
[INFO] [stdout]     |    ^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unreachable statement
[INFO] [stdout]   --> examples/3_modules_and_models.rs:40:5
[INFO] [stdout]    |
[INFO] [stdout] 38 |     return;
[INFO] [stdout]    |     ------ any code following this expression is unreachable
[INFO] [stdout] 39 |
[INFO] [stdout] 40 |     let linear_layer_config = LinearLayerConfig::new(3, 1);
[INFO] [stdout]    |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ unreachable statement
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(unreachable_code)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: variable does not need to be mutable
[INFO] [stdout]   --> examples/3_modules_and_models.rs:23:9
[INFO] [stdout]    |
[INFO] [stdout] 23 |     let mut tensor_from = Tensor::from_vec(vec![1.0, 2.0, 3.0,  4.0], vec![2, 2].into());
[INFO] [stdout]    |         ----^^^^^^^^^^^
[INFO] [stdout]    |         |
[INFO] [stdout]    |         help: remove this `mut`
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(unused_mut)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: variable does not need to be mutable
[INFO] [stdout]   --> examples/3_modules_and_models.rs:24:9
[INFO] [stdout]    |
[INFO] [stdout] 24 |     let mut bias = Tensor::from_vec(vec![0.0, 0.0], vec![2].into());
[INFO] [stdout]    |         ----^^^^
[INFO] [stdout]    |         |
[INFO] [stdout]    |         help: remove this `mut`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unreachable statement
[INFO] [stdout]   --> src/nn/layers/embedding.rs:31:9
[INFO] [stdout]    |
[INFO] [stdout] 29 |         return self.tensor.embbeding(input, self.model_dimension);
[INFO] [stdout]    |         --------------------------------------------------------- any code following this expression is unreachable
[INFO] [stdout] 30 |         
[INFO] [stdout] 31 |         let mut test_index_tensor = Tensor::zeroes(Shape::new(vec![input.shape.indices[0], input.shape.indices[1], self.model_dimension]));
[INFO] [stdout]    |         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ unreachable statement
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(unreachable_code)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unreachable statement
[INFO] [stdout]    --> src/nn/layers/layer_norm.rs:200:9
[INFO] [stdout]     |
[INFO] [stdout] 198 | ...   panic!("---");
[INFO] [stdout]     |       ------------- any code following this expression is unreachable
[INFO] [stdout] 199 | ...
[INFO] [stdout] 200 | ...   let expected_layer_norm_weight_grad_output_flatten = expected_layer_norm_weight_grad.item().iter().map(|x| x.clone()).collect::<Vec<f32>>();
[INFO] [stdout]     |       ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ unreachable statement
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `test_input`
[INFO] [stdout]    --> src/nn/layers/layer_norm.rs:157:13
[INFO] [stdout]     |
[INFO] [stdout] 157 |         let test_input = Tensor::from_bytestream(&mut test_input_file, false).unwrap();
[INFO] [stdout]     |             ^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_test_input`
[INFO] [stdout]     |
[INFO] [stdout]     = note: `#[warn(unused_variables)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `expected_layer_norm_weight_grad`
[INFO] [stdout]    --> src/nn/layers/layer_norm.rs:163:13
[INFO] [stdout]     |
[INFO] [stdout] 163 |         let expected_layer_norm_weight_grad = Tensor::from_bytestream(&mut layer_norm_weight_grad_file, false).unwrap();
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_expected_layer_norm_weight_grad`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `layer_norm_bias_grad`
[INFO] [stdout]    --> src/nn/layers/layer_norm.rs:164:13
[INFO] [stdout]     |
[INFO] [stdout] 164 |         let layer_norm_bias_grad = Tensor::from_bytestream(&mut layer_norm_bias_grad_file, false).unwrap();
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_layer_norm_bias_grad`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `o`
[INFO] [stdout]    --> src/nn/layers/layer_norm.rs:179:14
[INFO] [stdout]     |
[INFO] [stdout] 179 |         for (o, e) in ouput_as_flat_array.iter().zip(expected_as_flat_array.iter()) {
[INFO] [stdout]     |              ^ help: if this is intentional, prefix it with an underscore: `_o`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `e`
[INFO] [stdout]    --> src/nn/layers/layer_norm.rs:179:17
[INFO] [stdout]     |
[INFO] [stdout] 179 |         for (o, e) in ouput_as_flat_array.iter().zip(expected_as_flat_array.iter()) {
[INFO] [stdout]     |                 ^ help: if this is intentional, prefix it with an underscore: `_e`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `loss`
[INFO] [stdout]    --> src/nn/layers/layer_norm.rs:189:14
[INFO] [stdout]     |
[INFO] [stdout] 189 |         for (loss, expected) in mse_loss.item().iter().zip(expected_mse_loss.iter()) {
[INFO] [stdout]     |              ^^^^ help: if this is intentional, prefix it with an underscore: `_loss`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `expected`
[INFO] [stdout]    --> src/nn/layers/layer_norm.rs:189:20
[INFO] [stdout]     |
[INFO] [stdout] 189 |         for (loss, expected) in mse_loss.item().iter().zip(expected_mse_loss.iter()) {
[INFO] [stdout]     |                    ^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_expected`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `linear_layer`
[INFO] [stdout]   --> src/nn/layers/linear.rs:63:17
[INFO] [stdout]    |
[INFO] [stdout] 63 |         let mut linear_layer = LinearLayer::new(config);
[INFO] [stdout]    |                 ^^^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_linear_layer`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: variable does not need to be mutable
[INFO] [stdout]   --> src/nn/layers/linear.rs:63:13
[INFO] [stdout]    |
[INFO] [stdout] 63 |         let mut linear_layer = LinearLayer::new(config);
[INFO] [stdout]    |             ----^^^^^^^^^^^^
[INFO] [stdout]    |             |
[INFO] [stdout]    |             help: remove this `mut`
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(unused_mut)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `source_grad`
[INFO] [stdout]    --> src/central/equation.rs:805:21
[INFO] [stdout]     |
[INFO] [stdout] 805 |                 let source_grad = self.get_tensor_grad(origin_tensor).into_raw_vec();
[INFO] [stdout]     |                     ^^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_source_grad`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: field `vocab_size` is never read
[INFO] [stdout]  --> src/nn/layers/embedding.rs:7:5
[INFO] [stdout]   |
[INFO] [stdout] 5 | pub struct Embedding {
[INFO] [stdout]   |            --------- field in this struct
[INFO] [stdout] 6 |     pub tensor: Tensor,
[INFO] [stdout] 7 |     vocab_size: usize,
[INFO] [stdout]   |     ^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout]   = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: struct `MLPConfig` is never constructed
[INFO] [stdout]   --> src/nn/layers/mlp.rs:23:8
[INFO] [stdout]    |
[INFO] [stdout] 23 | struct MLPConfig {
[INFO] [stdout]    |        ^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: variable `M` should have a snake case name
[INFO] [stdout]    --> src/central/equation.rs:431:9
[INFO] [stdout]     |
[INFO] [stdout] 431 |     let M = a_shape[a_shape.len() - 2];
[INFO] [stdout]     |         ^ help: convert the identifier to snake case: `m`
[INFO] [stdout]     |
[INFO] [stdout]     = note: `#[warn(non_snake_case)]` (part of `#[warn(nonstandard_style)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: variable `K_a` should have a snake case name
[INFO] [stdout]    --> src/central/equation.rs:432:9
[INFO] [stdout]     |
[INFO] [stdout] 432 |     let K_a = a_shape[a_shape.len() - 1];
[INFO] [stdout]     |         ^^^ help: convert the identifier to snake case (notice the capitalization): `k_a`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `new` is never used
[INFO] [stdout]   --> src/nn/layers/mlp.rs:34:8
[INFO] [stdout]    |
[INFO] [stdout] 33 | impl MLP {
[INFO] [stdout]    | -------- associated function in this implementation
[INFO] [stdout] 34 |     fn new(config: MLPConfig) -> Self {
[INFO] [stdout]    |        ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: field `length_of_normalized_shape` is never read
[INFO] [stdout]  --> src/nn/layers/layer_norm.rs:9:5
[INFO] [stdout]   |
[INFO] [stdout] 6 | pub struct LayerNorm {
[INFO] [stdout]   |            --------- field in this struct
[INFO] [stdout] ...
[INFO] [stdout] 9 |     length_of_normalized_shape: usize,
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: variable `K_b` should have a snake case name
[INFO] [stdout]    --> src/central/equation.rs:433:9
[INFO] [stdout]     |
[INFO] [stdout] 433 |     let K_b = b_shape[b_shape.len() - 2];
[INFO] [stdout]     |         ^^^ help: convert the identifier to snake case (notice the capitalization): `k_b`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: variable `N` should have a snake case name
[INFO] [stdout]    --> src/central/equation.rs:434:9
[INFO] [stdout]     |
[INFO] [stdout] 434 |     let N = b_shape[b_shape.len() - 1];
[INFO] [stdout]     |         ^ help: convert the identifier to snake case: `n`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: method `add_time` is never used
[INFO] [stdout]    --> src/central/equation.rs:201:8
[INFO] [stdout]     |
[INFO] [stdout]  46 | impl Equation {
[INFO] [stdout]     | ------------- method in this implementation
[INFO] [stdout] ...
[INFO] [stdout] 201 |     fn add_time(&mut self, operation: &str, now: Instant) {
[INFO] [stdout]     |        ^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `write_f32_vector_to_file` is never used
[INFO] [stdout]  --> src/nn/layers/casual_self_attention.rs:7:4
[INFO] [stdout]   |
[INFO] [stdout] 7 | fn write_f32_vector_to_file(path: &str, data: &[f32]) -> std::io::Result<()> {
[INFO] [stdout]   |    ^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `write_string_vector_to_file` is never used
[INFO] [stdout]   --> src/nn/layers/casual_self_attention.rs:29:4
[INFO] [stdout]    |
[INFO] [stdout] 29 | fn write_string_vector_to_file(path: &str, data: &str) -> std::io::Result<()> {
[INFO] [stdout]    |    ^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: associated function `transpose_inside` is never used
[INFO] [stdout]    --> src/central/tensor.rs:632:8
[INFO] [stdout]     |
[INFO] [stdout]  45 | impl Tensor {
[INFO] [stdout]     | ----------- associated function in this implementation
[INFO] [stdout] ...
[INFO] [stdout] 632 |     fn transpose_inside(matrix: Vec<f32>, rows: usize, cols: usize) -> Vec<f32> {
[INFO] [stdout]     |        ^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stderr]     Finished `test` profile [unoptimized + debuginfo] target(s) in 6.55s
[INFO] running `Command { std: "docker" "inspect" "7edaba70dea76f816c8a4506f2d5e176c9b84f2c72c1fdba07bbe12cb7796183", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "7edaba70dea76f816c8a4506f2d5e176c9b84f2c72c1fdba07bbe12cb7796183", kill_on_drop: false }`
[INFO] [stdout] 7edaba70dea76f816c8a4506f2d5e176c9b84f2c72c1fdba07bbe12cb7796183
