[INFO] cloning repository https://github.com/TuckerBMorgan/hermes [INFO] running `Command { std: "git" "-c" "credential.helper=" "-c" "credential.helper=/workspace/cargo-home/bin/git-credential-null" "clone" "--bare" "https://github.com/TuckerBMorgan/hermes" "/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2FTuckerBMorgan%2Fhermes", kill_on_drop: false }` [INFO] [stderr] Cloning into bare repository '/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2FTuckerBMorgan%2Fhermes'... [INFO] running `Command { std: "git" "rev-parse" "HEAD", kill_on_drop: false }` [INFO] [stdout] c08cc264a7d7f3a75373d1b5efd7bb440de78131 [INFO] building TuckerBMorgan/hermes against try#d79d87d4c0f7f0d0785ad6e493cf8cf718ed89a6 for pr-146237-1 [INFO] running `Command { std: "git" "clone" "/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2FTuckerBMorgan%2Fhermes" "/workspace/builds/worker-7-tc2/source", kill_on_drop: false }` [INFO] [stderr] Cloning into '/workspace/builds/worker-7-tc2/source'... [INFO] [stderr] done. [INFO] started tweaking git repo https://github.com/TuckerBMorgan/hermes [INFO] finished tweaking git repo https://github.com/TuckerBMorgan/hermes [INFO] tweaked toml for git repo https://github.com/TuckerBMorgan/hermes written to /workspace/builds/worker-7-tc2/source/Cargo.toml [INFO] validating manifest of git repo https://github.com/TuckerBMorgan/hermes on toolchain d79d87d4c0f7f0d0785ad6e493cf8cf718ed89a6 [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+d79d87d4c0f7f0d0785ad6e493cf8cf718ed89a6" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }` [INFO] crate git repo https://github.com/TuckerBMorgan/hermes already has a lockfile, it will not be regenerated [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+d79d87d4c0f7f0d0785ad6e493cf8cf718ed89a6" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }` [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-7-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-7-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:e90291280db7d1fac5b66fc6dad9f9662629e7365a55743daf9bdf73ebc4ea79" "/opt/rustwide/cargo-home/bin/cargo" "+d79d87d4c0f7f0d0785ad6e493cf8cf718ed89a6" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }` [INFO] [stdout] 08dad828e1fff56ce468794ae0d51522052b24d94f91f69931a28b73986b8cf7 [INFO] running `Command { std: "docker" "start" "-a" "08dad828e1fff56ce468794ae0d51522052b24d94f91f69931a28b73986b8cf7", kill_on_drop: false }` [INFO] running `Command { std: "docker" "inspect" "08dad828e1fff56ce468794ae0d51522052b24d94f91f69931a28b73986b8cf7", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "08dad828e1fff56ce468794ae0d51522052b24d94f91f69931a28b73986b8cf7", kill_on_drop: false }` [INFO] [stdout] 08dad828e1fff56ce468794ae0d51522052b24d94f91f69931a28b73986b8cf7 [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-7-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-7-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:e90291280db7d1fac5b66fc6dad9f9662629e7365a55743daf9bdf73ebc4ea79" "/opt/rustwide/cargo-home/bin/cargo" "+d79d87d4c0f7f0d0785ad6e493cf8cf718ed89a6" "build" "--frozen" "--message-format=json", kill_on_drop: false }` [INFO] [stdout] 3f4afb8717afa997a55d28f1fb6eaa47a722618a31939e13f2648cc76cdae95b [INFO] running `Command { std: "docker" "start" "-a" "3f4afb8717afa997a55d28f1fb6eaa47a722618a31939e13f2648cc76cdae95b", kill_on_drop: false }` [INFO] [stderr] Compiling cc v1.2.7 [INFO] [stderr] Compiling num-traits v0.2.19 [INFO] [stderr] Compiling smallvec v1.13.2 [INFO] [stderr] Compiling either v1.13.0 [INFO] [stderr] Compiling openssl v0.10.68 [INFO] [stderr] Compiling syn v2.0.95 [INFO] [stderr] Compiling portable-atomic v1.10.0 [INFO] [stderr] Compiling rustls v0.23.20 [INFO] [stderr] Compiling percent-encoding v2.3.1 [INFO] [stderr] Compiling unicode-width v0.2.0 [INFO] [stderr] Compiling dirs-sys v0.4.1 [INFO] [stderr] Compiling webpki-roots v0.26.7 [INFO] [stderr] Compiling matrixmultiply v0.3.9 [INFO] [stderr] Compiling aho-corasick v1.1.3 [INFO] [stderr] Compiling number_prefix v0.4.0 [INFO] [stderr] Compiling macro_rules_attribute-proc_macro v0.2.0 [INFO] [stderr] Compiling unicode-segmentation v1.12.0 [INFO] [stderr] Compiling unicode-normalization-alignments v0.1.12 [INFO] [stderr] Compiling termcolor v1.1.3 [INFO] [stderr] Compiling form_urlencoded v1.2.1 [INFO] [stderr] Compiling rayon v1.10.0 [INFO] [stderr] Compiling console v0.15.10 [INFO] [stderr] Compiling dirs v5.0.1 [INFO] [stderr] Compiling itertools v0.11.0 [INFO] [stderr] Compiling itertools v0.12.1 [INFO] [stderr] Compiling macro_rules_attribute v0.2.0 [INFO] [stderr] Compiling num-integer v0.1.46 [INFO] [stderr] Compiling num-complex v0.4.6 [INFO] [stderr] Compiling chrono v0.4.39 [INFO] [stderr] Compiling regex-automata v0.4.9 [INFO] [stderr] Compiling indicatif v0.17.9 [INFO] [stderr] Compiling openssl-sys v0.9.104 [INFO] [stderr] Compiling ring v0.17.8 [INFO] [stderr] Compiling onig_sys v69.8.1 [INFO] [stderr] Compiling esaxx-rs v0.1.10 [INFO] [stderr] Compiling rayon-cond v0.3.0 [INFO] [stderr] Compiling native-tls v0.2.12 [INFO] [stderr] Compiling ndarray v0.15.6 [INFO] [stderr] Compiling simplelog v0.11.2 [INFO] [stderr] Compiling synstructure v0.13.1 [INFO] [stderr] Compiling darling_core v0.20.10 [INFO] [stderr] Compiling regex v1.11.1 [INFO] [stderr] Compiling zerovec-derive v0.10.3 [INFO] [stderr] Compiling zerofrom-derive v0.1.5 [INFO] [stderr] Compiling yoke-derive v0.7.5 [INFO] [stderr] Compiling displaydoc v0.2.5 [INFO] [stderr] Compiling icu_provider_macros v1.5.0 [INFO] [stderr] Compiling serde_derive v1.0.217 [INFO] [stderr] Compiling zerocopy-derive v0.7.35 [INFO] [stderr] Compiling openssl-macros v0.1.1 [INFO] [stderr] Compiling thiserror-impl v1.0.69 [INFO] [stderr] Compiling monostate-impl v0.1.13 [INFO] [stderr] Compiling zerocopy v0.7.35 [INFO] [stderr] Compiling zerofrom v0.1.5 [INFO] [stderr] Compiling yoke v0.7.5 [INFO] [stderr] Compiling zerovec v0.10.4 [INFO] [stderr] Compiling ppv-lite86 v0.2.20 [INFO] [stderr] Compiling rustls-webpki v0.102.8 [INFO] [stderr] Compiling darling_macro v0.20.10 [INFO] [stderr] Compiling thiserror v1.0.69 [INFO] [stderr] Compiling rand_chacha v0.3.1 [INFO] [stderr] Compiling onig v6.4.0 [INFO] [stderr] Compiling darling v0.20.10 [INFO] [stderr] Compiling derive_builder_core v0.20.2 [INFO] [stderr] Compiling tinystr v0.7.6 [INFO] [stderr] Compiling icu_collections v1.5.0 [INFO] [stderr] Compiling rand v0.8.5 [INFO] [stderr] Compiling icu_locid v1.5.0 [INFO] [stderr] Compiling icu_provider v1.5.0 [INFO] [stderr] Compiling serde v1.0.217 [INFO] [stderr] Compiling rand_distr v0.4.3 [INFO] [stderr] Compiling icu_locid_transform v1.5.0 [INFO] [stderr] Compiling derive_builder_macro v0.20.2 [INFO] [stderr] Compiling icu_properties v1.5.1 [INFO] [stderr] Compiling derive_builder v0.20.2 [INFO] [stderr] Compiling icu_normalizer v1.5.0 [INFO] [stderr] Compiling serde_json v1.0.135 [INFO] [stderr] Compiling monostate v0.1.13 [INFO] [stderr] Compiling spm_precompiled v0.1.4 [INFO] [stderr] Compiling idna_adapter v1.2.0 [INFO] [stderr] Compiling idna v1.0.3 [INFO] [stderr] Compiling url v2.5.4 [INFO] [stderr] Compiling ureq v2.12.1 [INFO] [stderr] Compiling hf-hub v0.3.2 [INFO] [stderr] Compiling tokenizers v0.20.4 [INFO] [stderr] Compiling poro v0.1.2 (/opt/rustwide/workdir) [INFO] [stdout] warning: unused import: `ndarray::prelude::*` [INFO] [stdout] --> src/central/add_op.rs:7:9 [INFO] [stdout] | [INFO] [stdout] 7 | pub use ndarray::prelude::*; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `Result` [INFO] [stdout] --> src/central/tensor.rs:7:21 [INFO] [stdout] | [INFO] [stdout] 7 | use std::io::{Read, Result}; [INFO] [stdout] | ^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unnecessary trailing semicolon [INFO] [stdout] --> src/central/tensor.rs:224:92 [INFO] [stdout] | [INFO] [stdout] 224 | .map(|chunk| i32::from_le_bytes(chunk.try_into().unwrap()) as usize).collect();; [INFO] [stdout] | ^ help: remove this semicolon [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(redundant_semicolons)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused imports: `LinearLayerConfig` and `LinearLayer` [INFO] [stdout] --> src/nn/layers/linear.rs:55:37 [INFO] [stdout] | [INFO] [stdout] 55 | use crate::nn::layers::linear::{LinearLayer, LinearLayerConfig}; [INFO] [stdout] | ^^^^^^^^^^^ ^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::nn::layers::module::Module` [INFO] [stdout] --> src/nn/layers/linear.rs:56:9 [INFO] [stdout] | [INFO] [stdout] 56 | use crate::nn::layers::module::Module; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::central::Tensor` [INFO] [stdout] --> src/nn/layers/linear.rs:57:9 [INFO] [stdout] | [INFO] [stdout] 57 | use crate::central::Tensor; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::Shape` [INFO] [stdout] --> src/nn/layers/linear.rs:58:9 [INFO] [stdout] | [INFO] [stdout] 58 | use crate::Shape; [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::nn::layers::linear` [INFO] [stdout] --> src/nn/layers/mlp.rs:80:9 [INFO] [stdout] | [INFO] [stdout] 80 | use crate::nn::layers::linear; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::nn::layers::mlp::MLP` [INFO] [stdout] --> src/nn/layers/mlp.rs:81:9 [INFO] [stdout] | [INFO] [stdout] 81 | use crate::nn::layers::mlp::MLP; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::nn::layers::module::Module` [INFO] [stdout] --> src/nn/layers/mlp.rs:82:9 [INFO] [stdout] | [INFO] [stdout] 82 | use crate::nn::layers::module::Module; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::central::Tensor` [INFO] [stdout] --> src/nn/layers/mlp.rs:83:9 [INFO] [stdout] | [INFO] [stdout] 83 | use crate::central::Tensor; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::Shape` [INFO] [stdout] --> src/nn/layers/mlp.rs:84:9 [INFO] [stdout] | [INFO] [stdout] 84 | use crate::Shape; [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `core::panic` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:1:5 [INFO] [stdout] | [INFO] [stdout] 1 | use core::panic; [INFO] [stdout] | ^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unnecessary trailing semicolon [INFO] [stdout] --> src/nn/layers/layer_norm.rs:56:47 [INFO] [stdout] | [INFO] [stdout] 56 | let var: Tensor = x.var(mean_indices);;//(input_minus_mean - input_minus_mean).pow(2.0).mean(mean_indices.clone()); [INFO] [stdout] | ^ help: remove this semicolon [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `core::panic` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:77:9 [INFO] [stdout] | [INFO] [stdout] 77 | use core::panic; [INFO] [stdout] | ^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `ndarray::prelude::*` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:79:9 [INFO] [stdout] | [INFO] [stdout] 79 | use ndarray::prelude::*; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `ndarray::iter::Axes` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:87:9 [INFO] [stdout] | [INFO] [stdout] 87 | use ndarray::iter::Axes; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `ndarray::Axis` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:88:9 [INFO] [stdout] | [INFO] [stdout] 88 | use ndarray::Axis; [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::central::Tensor` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:90:9 [INFO] [stdout] | [INFO] [stdout] 90 | use crate::central::Tensor; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::nn::layers::layer_norm::LayerNorm` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:91:9 [INFO] [stdout] | [INFO] [stdout] 91 | use crate::nn::layers::layer_norm::LayerNorm; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::nn::layers::module::Module` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:92:9 [INFO] [stdout] | [INFO] [stdout] 92 | use crate::nn::layers::module::Module; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `simplelog::*` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:93:9 [INFO] [stdout] | [INFO] [stdout] 93 | use simplelog::*; [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `log::info` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:94:9 [INFO] [stdout] | [INFO] [stdout] 94 | use log::info; [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `std::path::Path` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:95:9 [INFO] [stdout] | [INFO] [stdout] 95 | use std::path::Path; [INFO] [stdout] | ^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `std::fs::File` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:96:9 [INFO] [stdout] | [INFO] [stdout] 96 | use std::fs::File; [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `std::vec` [INFO] [stdout] --> src/nn/layers/casual_self_attention.rs:203:9 [INFO] [stdout] | [INFO] [stdout] 203 | use std::vec; [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::nn::layers::linear::LinearLayerConfig` [INFO] [stdout] --> src/nn/layers/casual_self_attention.rs:204:9 [INFO] [stdout] | [INFO] [stdout] 204 | use crate::nn::layers::linear::LinearLayerConfig; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::nn::layers::module::Module` [INFO] [stdout] --> src/nn/layers/casual_self_attention.rs:205:9 [INFO] [stdout] | [INFO] [stdout] 205 | use crate::nn::layers::module::Module; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::central::Tensor` [INFO] [stdout] --> src/nn/layers/casual_self_attention.rs:206:9 [INFO] [stdout] | [INFO] [stdout] 206 | use crate::central::Tensor; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::Shape` [INFO] [stdout] --> src/nn/layers/casual_self_attention.rs:207:9 [INFO] [stdout] | [INFO] [stdout] 207 | use crate::Shape; [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `std::hash::Hash` [INFO] [stdout] --> src/optimizers/adam.rs:2:5 [INFO] [stdout] | [INFO] [stdout] 2 | use std::hash::Hash; [INFO] [stdout] | ^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `std::hash::Hash` [INFO] [stdout] --> src/optimizers/adamw.rs:2:5 [INFO] [stdout] | [INFO] [stdout] 2 | use std::hash::Hash; [INFO] [stdout] | ^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unreachable statement [INFO] [stdout] --> src/nn/layers/embedding.rs:31:9 [INFO] [stdout] | [INFO] [stdout] 29 | return self.tensor.embbeding(input, self.model_dimension); [INFO] [stdout] | --------------------------------------------------------- any code following this expression is unreachable [INFO] [stdout] 30 | [INFO] [stdout] 31 | let mut test_index_tensor = Tensor::zeroes(Shape::new(vec![input.shape.indices[0], input.shape.indices[1], self.model_dimension])); [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ unreachable statement [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unreachable_code)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `source_grad` [INFO] [stdout] --> src/central/equation.rs:805:21 [INFO] [stdout] | [INFO] [stdout] 805 | let source_grad = self.get_tensor_grad(origin_tensor).into_raw_vec(); [INFO] [stdout] | ^^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_source_grad` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_variables)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variable `M` should have a snake case name [INFO] [stdout] --> src/central/equation.rs:431:9 [INFO] [stdout] | [INFO] [stdout] 431 | let M = a_shape[a_shape.len() - 2]; [INFO] [stdout] | ^ help: convert the identifier to snake case: `m` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(non_snake_case)]` (part of `#[warn(nonstandard_style)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variable `K_a` should have a snake case name [INFO] [stdout] --> src/central/equation.rs:432:9 [INFO] [stdout] | [INFO] [stdout] 432 | let K_a = a_shape[a_shape.len() - 1]; [INFO] [stdout] | ^^^ help: convert the identifier to snake case (notice the capitalization): `k_a` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variable `K_b` should have a snake case name [INFO] [stdout] --> src/central/equation.rs:433:9 [INFO] [stdout] | [INFO] [stdout] 433 | let K_b = b_shape[b_shape.len() - 2]; [INFO] [stdout] | ^^^ help: convert the identifier to snake case (notice the capitalization): `k_b` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variable `N` should have a snake case name [INFO] [stdout] --> src/central/equation.rs:434:9 [INFO] [stdout] | [INFO] [stdout] 434 | let N = b_shape[b_shape.len() - 1]; [INFO] [stdout] | ^ help: convert the identifier to snake case: `n` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: method `add_time` is never used [INFO] [stdout] --> src/central/equation.rs:201:8 [INFO] [stdout] | [INFO] [stdout] 46 | impl Equation { [INFO] [stdout] | ------------- method in this implementation [INFO] [stdout] ... [INFO] [stdout] 201 | fn add_time(&mut self, operation: &str, now: Instant) { [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: associated function `transpose_inside` is never used [INFO] [stdout] --> src/central/tensor.rs:632:8 [INFO] [stdout] | [INFO] [stdout] 45 | impl Tensor { [INFO] [stdout] | ----------- associated function in this implementation [INFO] [stdout] ... [INFO] [stdout] 632 | fn transpose_inside(matrix: Vec, rows: usize, cols: usize) -> Vec { [INFO] [stdout] | ^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: struct `MLPConfig` is never constructed [INFO] [stdout] --> src/nn/layers/mlp.rs:23:8 [INFO] [stdout] | [INFO] [stdout] 23 | struct MLPConfig { [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: associated function `new` is never used [INFO] [stdout] --> src/nn/layers/mlp.rs:34:8 [INFO] [stdout] | [INFO] [stdout] 33 | impl MLP { [INFO] [stdout] | -------- associated function in this implementation [INFO] [stdout] 34 | fn new(config: MLPConfig) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `length_of_normalized_shape` is never read [INFO] [stdout] --> src/nn/layers/layer_norm.rs:9:5 [INFO] [stdout] | [INFO] [stdout] 6 | pub struct LayerNorm { [INFO] [stdout] | --------- field in this struct [INFO] [stdout] ... [INFO] [stdout] 9 | length_of_normalized_shape: usize, [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `write_f32_vector_to_file` is never used [INFO] [stdout] --> src/nn/layers/casual_self_attention.rs:7:4 [INFO] [stdout] | [INFO] [stdout] 7 | fn write_f32_vector_to_file(path: &str, data: &[f32]) -> std::io::Result<()> { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `write_string_vector_to_file` is never used [INFO] [stdout] --> src/nn/layers/casual_self_attention.rs:29:4 [INFO] [stdout] | [INFO] [stdout] 29 | fn write_string_vector_to_file(path: &str, data: &str) -> std::io::Result<()> { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `vocab_size` is never read [INFO] [stdout] --> src/nn/layers/embedding.rs:7:5 [INFO] [stdout] | [INFO] [stdout] 5 | pub struct Embedding { [INFO] [stdout] | --------- field in this struct [INFO] [stdout] 6 | pub tensor: Tensor, [INFO] [stdout] 7 | vocab_size: usize, [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] Finished `dev` profile [unoptimized + debuginfo] target(s) in 1m 15s [INFO] running `Command { std: "docker" "inspect" "3f4afb8717afa997a55d28f1fb6eaa47a722618a31939e13f2648cc76cdae95b", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "3f4afb8717afa997a55d28f1fb6eaa47a722618a31939e13f2648cc76cdae95b", kill_on_drop: false }` [INFO] [stdout] 3f4afb8717afa997a55d28f1fb6eaa47a722618a31939e13f2648cc76cdae95b [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-7-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-7-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:e90291280db7d1fac5b66fc6dad9f9662629e7365a55743daf9bdf73ebc4ea79" "/opt/rustwide/cargo-home/bin/cargo" "+d79d87d4c0f7f0d0785ad6e493cf8cf718ed89a6" "test" "--frozen" "--no-run" "--message-format=json", kill_on_drop: false }` [INFO] [stdout] 7edaba70dea76f816c8a4506f2d5e176c9b84f2c72c1fdba07bbe12cb7796183 [INFO] running `Command { std: "docker" "start" "-a" "7edaba70dea76f816c8a4506f2d5e176c9b84f2c72c1fdba07bbe12cb7796183", kill_on_drop: false }` [INFO] [stdout] warning: unused import: `ndarray::prelude::*` [INFO] [stdout] --> src/central/add_op.rs:7:9 [INFO] [stdout] | [INFO] [stdout] 7 | pub use ndarray::prelude::*; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `Result` [INFO] [stdout] --> src/central/tensor.rs:7:21 [INFO] [stdout] | [INFO] [stdout] 7 | use std::io::{Read, Result}; [INFO] [stdout] | ^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unnecessary trailing semicolon [INFO] [stdout] --> src/central/tensor.rs:224:92 [INFO] [stdout] | [INFO] [stdout] 224 | .map(|chunk| i32::from_le_bytes(chunk.try_into().unwrap()) as usize).collect();; [INFO] [stdout] | ^ help: remove this semicolon [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(redundant_semicolons)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused imports: `LinearLayerConfig` and `LinearLayer` [INFO] [stdout] --> src/nn/layers/linear.rs:55:37 [INFO] [stdout] | [INFO] [stdout] 55 | use crate::nn::layers::linear::{LinearLayer, LinearLayerConfig}; [INFO] [stdout] | ^^^^^^^^^^^ ^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::nn::layers::module::Module` [INFO] [stdout] --> src/nn/layers/linear.rs:56:9 [INFO] [stdout] | [INFO] [stdout] 56 | use crate::nn::layers::module::Module; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::central::Tensor` [INFO] [stdout] --> src/nn/layers/linear.rs:57:9 [INFO] [stdout] | [INFO] [stdout] 57 | use crate::central::Tensor; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::Shape` [INFO] [stdout] --> src/nn/layers/linear.rs:58:9 [INFO] [stdout] | [INFO] [stdout] 58 | use crate::Shape; [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::nn::layers::linear` [INFO] [stdout] --> src/nn/layers/mlp.rs:80:9 [INFO] [stdout] | [INFO] [stdout] 80 | use crate::nn::layers::linear; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::nn::layers::mlp::MLP` [INFO] [stdout] --> src/nn/layers/mlp.rs:81:9 [INFO] [stdout] | [INFO] [stdout] 81 | use crate::nn::layers::mlp::MLP; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::nn::layers::module::Module` [INFO] [stdout] --> src/nn/layers/mlp.rs:82:9 [INFO] [stdout] | [INFO] [stdout] 82 | use crate::nn::layers::module::Module; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::central::Tensor` [INFO] [stdout] --> src/nn/layers/mlp.rs:83:9 [INFO] [stdout] | [INFO] [stdout] 83 | use crate::central::Tensor; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::Shape` [INFO] [stdout] --> src/nn/layers/mlp.rs:84:9 [INFO] [stdout] | [INFO] [stdout] 84 | use crate::Shape; [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `core::panic` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:1:5 [INFO] [stdout] | [INFO] [stdout] 1 | use core::panic; [INFO] [stdout] | ^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unnecessary trailing semicolon [INFO] [stdout] --> src/nn/layers/layer_norm.rs:56:47 [INFO] [stdout] | [INFO] [stdout] 56 | let var: Tensor = x.var(mean_indices);;//(input_minus_mean - input_minus_mean).pow(2.0).mean(mean_indices.clone()); [INFO] [stdout] | ^ help: remove this semicolon [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `core::panic` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:77:9 [INFO] [stdout] | [INFO] [stdout] 77 | use core::panic; [INFO] [stdout] | ^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `ndarray::prelude::*` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:79:9 [INFO] [stdout] | [INFO] [stdout] 79 | use ndarray::prelude::*; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `ndarray::iter::Axes` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:87:9 [INFO] [stdout] | [INFO] [stdout] 87 | use ndarray::iter::Axes; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `ndarray::Axis` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:88:9 [INFO] [stdout] | [INFO] [stdout] 88 | use ndarray::Axis; [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::central::Tensor` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:90:9 [INFO] [stdout] | [INFO] [stdout] 90 | use crate::central::Tensor; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::nn::layers::layer_norm::LayerNorm` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:91:9 [INFO] [stdout] | [INFO] [stdout] 91 | use crate::nn::layers::layer_norm::LayerNorm; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::nn::layers::module::Module` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:92:9 [INFO] [stdout] | [INFO] [stdout] 92 | use crate::nn::layers::module::Module; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `simplelog::*` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:93:9 [INFO] [stdout] | [INFO] [stdout] 93 | use simplelog::*; [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `log::info` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:94:9 [INFO] [stdout] | [INFO] [stdout] 94 | use log::info; [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `std::path::Path` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:95:9 [INFO] [stdout] | [INFO] [stdout] 95 | use std::path::Path; [INFO] [stdout] | ^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `std::fs::File` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:96:9 [INFO] [stdout] | [INFO] [stdout] 96 | use std::fs::File; [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `std::vec` [INFO] [stdout] --> src/nn/layers/casual_self_attention.rs:203:9 [INFO] [stdout] | [INFO] [stdout] 203 | use std::vec; [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::nn::layers::linear::LinearLayerConfig` [INFO] [stdout] --> src/nn/layers/casual_self_attention.rs:204:9 [INFO] [stdout] | [INFO] [stdout] 204 | use crate::nn::layers::linear::LinearLayerConfig; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::nn::layers::module::Module` [INFO] [stdout] --> src/nn/layers/casual_self_attention.rs:205:9 [INFO] [stdout] | [INFO] [stdout] 205 | use crate::nn::layers::module::Module; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::central::Tensor` [INFO] [stdout] --> src/nn/layers/casual_self_attention.rs:206:9 [INFO] [stdout] | [INFO] [stdout] 206 | use crate::central::Tensor; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::Shape` [INFO] [stdout] --> src/nn/layers/casual_self_attention.rs:207:9 [INFO] [stdout] | [INFO] [stdout] 207 | use crate::Shape; [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `std::hash::Hash` [INFO] [stdout] --> src/optimizers/adam.rs:2:5 [INFO] [stdout] | [INFO] [stdout] 2 | use std::hash::Hash; [INFO] [stdout] | ^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `std::hash::Hash` [INFO] [stdout] --> src/optimizers/adamw.rs:2:5 [INFO] [stdout] | [INFO] [stdout] 2 | use std::hash::Hash; [INFO] [stdout] | ^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unreachable statement [INFO] [stdout] --> src/nn/layers/embedding.rs:31:9 [INFO] [stdout] | [INFO] [stdout] 29 | return self.tensor.embbeding(input, self.model_dimension); [INFO] [stdout] | --------------------------------------------------------- any code following this expression is unreachable [INFO] [stdout] 30 | [INFO] [stdout] 31 | let mut test_index_tensor = Tensor::zeroes(Shape::new(vec![input.shape.indices[0], input.shape.indices[1], self.model_dimension])); [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ unreachable statement [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unreachable_code)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `source_grad` [INFO] [stdout] --> src/central/equation.rs:805:21 [INFO] [stdout] | [INFO] [stdout] 805 | let source_grad = self.get_tensor_grad(origin_tensor).into_raw_vec(); [INFO] [stdout] | ^^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_source_grad` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_variables)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variable `M` should have a snake case name [INFO] [stdout] --> src/central/equation.rs:431:9 [INFO] [stdout] | [INFO] [stdout] 431 | let M = a_shape[a_shape.len() - 2]; [INFO] [stdout] | ^ help: convert the identifier to snake case: `m` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(non_snake_case)]` (part of `#[warn(nonstandard_style)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] Compiling poro v0.1.2 (/opt/rustwide/workdir) [INFO] [stdout] warning: variable `K_a` should have a snake case name [INFO] [stdout] --> src/central/equation.rs:432:9 [INFO] [stdout] | [INFO] [stdout] 432 | let K_a = a_shape[a_shape.len() - 1]; [INFO] [stdout] | ^^^ help: convert the identifier to snake case (notice the capitalization): `k_a` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variable `K_b` should have a snake case name [INFO] [stdout] --> src/central/equation.rs:433:9 [INFO] [stdout] | [INFO] [stdout] 433 | let K_b = b_shape[b_shape.len() - 2]; [INFO] [stdout] | ^^^ help: convert the identifier to snake case (notice the capitalization): `k_b` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variable `N` should have a snake case name [INFO] [stdout] --> src/central/equation.rs:434:9 [INFO] [stdout] | [INFO] [stdout] 434 | let N = b_shape[b_shape.len() - 1]; [INFO] [stdout] | ^ help: convert the identifier to snake case: `n` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: method `add_time` is never used [INFO] [stdout] --> src/central/equation.rs:201:8 [INFO] [stdout] | [INFO] [stdout] 46 | impl Equation { [INFO] [stdout] | ------------- method in this implementation [INFO] [stdout] ... [INFO] [stdout] 201 | fn add_time(&mut self, operation: &str, now: Instant) { [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: associated function `transpose_inside` is never used [INFO] [stdout] --> src/central/tensor.rs:632:8 [INFO] [stdout] | [INFO] [stdout] 45 | impl Tensor { [INFO] [stdout] | ----------- associated function in this implementation [INFO] [stdout] ... [INFO] [stdout] 632 | fn transpose_inside(matrix: Vec, rows: usize, cols: usize) -> Vec { [INFO] [stdout] | ^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: struct `MLPConfig` is never constructed [INFO] [stdout] --> src/nn/layers/mlp.rs:23:8 [INFO] [stdout] | [INFO] [stdout] 23 | struct MLPConfig { [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: associated function `new` is never used [INFO] [stdout] --> src/nn/layers/mlp.rs:34:8 [INFO] [stdout] | [INFO] [stdout] 33 | impl MLP { [INFO] [stdout] | -------- associated function in this implementation [INFO] [stdout] 34 | fn new(config: MLPConfig) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `length_of_normalized_shape` is never read [INFO] [stdout] --> src/nn/layers/layer_norm.rs:9:5 [INFO] [stdout] | [INFO] [stdout] 6 | pub struct LayerNorm { [INFO] [stdout] | --------- field in this struct [INFO] [stdout] ... [INFO] [stdout] 9 | length_of_normalized_shape: usize, [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `write_f32_vector_to_file` is never used [INFO] [stdout] --> src/nn/layers/casual_self_attention.rs:7:4 [INFO] [stdout] | [INFO] [stdout] 7 | fn write_f32_vector_to_file(path: &str, data: &[f32]) -> std::io::Result<()> { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `write_string_vector_to_file` is never used [INFO] [stdout] --> src/nn/layers/casual_self_attention.rs:29:4 [INFO] [stdout] | [INFO] [stdout] 29 | fn write_string_vector_to_file(path: &str, data: &str) -> std::io::Result<()> { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `vocab_size` is never read [INFO] [stdout] --> src/nn/layers/embedding.rs:7:5 [INFO] [stdout] | [INFO] [stdout] 5 | pub struct Embedding { [INFO] [stdout] | --------- field in this struct [INFO] [stdout] 6 | pub tensor: Tensor, [INFO] [stdout] 7 | vocab_size: usize, [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: associated function `from_tensor` is never used [INFO] [stdout] --> examples/wavenet.rs:20:12 [INFO] [stdout] | [INFO] [stdout] 13 | impl EmbeddingLayer { [INFO] [stdout] | ------------------- associated function in this implementation [INFO] [stdout] ... [INFO] [stdout] 20 | pub fn from_tensor(weights: Tensor) -> Self { [INFO] [stdout] | ^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused imports: `AdamBuilder`, `Adam`, and `LinearLayerConfig` [INFO] [stdout] --> examples/linear_and_adam_optimizer.rs:1:44 [INFO] [stdout] | [INFO] [stdout] 1 | use poro::{get_equation, nn::{LinearLayer, LinearLayerConfig}, Adam, AdamBuilder, AdamW, AdamWBuilder, Optimizer}; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^ ^^^^ ^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `Result` [INFO] [stdout] --> src/central/tensor.rs:7:21 [INFO] [stdout] | [INFO] [stdout] 7 | use std::io::{Read, Result}; [INFO] [stdout] | ^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unnecessary trailing semicolon [INFO] [stdout] --> src/central/tensor.rs:224:92 [INFO] [stdout] | [INFO] [stdout] 224 | .map(|chunk| i32::from_le_bytes(chunk.try_into().unwrap()) as usize).collect();; [INFO] [stdout] | ^ help: remove this semicolon [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(redundant_semicolons)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::Shape` [INFO] [stdout] --> src/nn/layers/linear.rs:58:9 [INFO] [stdout] | [INFO] [stdout] 58 | use crate::Shape; [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::nn::layers::linear` [INFO] [stdout] --> src/nn/layers/mlp.rs:80:9 [INFO] [stdout] | [INFO] [stdout] 80 | use crate::nn::layers::linear; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::Shape` [INFO] [stdout] --> src/nn/layers/mlp.rs:84:9 [INFO] [stdout] | [INFO] [stdout] 84 | use crate::Shape; [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `core::panic` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:1:5 [INFO] [stdout] | [INFO] [stdout] 1 | use core::panic; [INFO] [stdout] | ^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unnecessary trailing semicolon [INFO] [stdout] --> src/nn/layers/layer_norm.rs:56:47 [INFO] [stdout] | [INFO] [stdout] 56 | let var: Tensor = x.var(mean_indices);;//(input_minus_mean - input_minus_mean).pow(2.0).mean(mean_indices.clone()); [INFO] [stdout] | ^ help: remove this semicolon [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `ndarray::iter::Axes` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:87:9 [INFO] [stdout] | [INFO] [stdout] 87 | use ndarray::iter::Axes; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `ndarray::Axis` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:88:9 [INFO] [stdout] | [INFO] [stdout] 88 | use ndarray::Axis; [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `log::info` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:94:9 [INFO] [stdout] | [INFO] [stdout] 94 | use log::info; [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::nn::layers::linear::LinearLayerConfig` [INFO] [stdout] --> src/nn/layers/casual_self_attention.rs:204:9 [INFO] [stdout] | [INFO] [stdout] 204 | use crate::nn::layers::linear::LinearLayerConfig; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::central::Tensor` [INFO] [stdout] --> src/nn/layers/casual_self_attention.rs:206:9 [INFO] [stdout] | [INFO] [stdout] 206 | use crate::central::Tensor; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::Shape` [INFO] [stdout] --> src/nn/layers/casual_self_attention.rs:207:9 [INFO] [stdout] | [INFO] [stdout] 207 | use crate::Shape; [INFO] [stdout] | ^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `std::hash::Hash` [INFO] [stdout] --> src/optimizers/adam.rs:2:5 [INFO] [stdout] | [INFO] [stdout] 2 | use std::hash::Hash; [INFO] [stdout] | ^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `std::hash::Hash` [INFO] [stdout] --> src/optimizers/adamw.rs:2:5 [INFO] [stdout] | [INFO] [stdout] 2 | use std::hash::Hash; [INFO] [stdout] | ^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `CasualSelfAttentionConfig` [INFO] [stdout] --> examples/transfomer_test.rs:4:107 [INFO] [stdout] | [INFO] [stdout] 4 | use poro::nn::{AttentionHead, Embedding, LinearLayer, Model, Module, MLP, LayerNorm, CasualSelfAttention, CasualSelfAttentionConfig}; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `poro::optimizers::adam::Adam` [INFO] [stdout] --> examples/transfomer_test.rs:6:5 [INFO] [stdout] | [INFO] [stdout] 6 | use poro::optimizers::adam::Adam; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `poro::nn::layers::NewGLU` [INFO] [stdout] --> examples/transfomer_test.rs:7:5 [INFO] [stdout] | [INFO] [stdout] 7 | use poro::nn::layers::NewGLU; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `Seek` [INFO] [stdout] --> examples/transfomer_test.rs:8:21 [INFO] [stdout] | [INFO] [stdout] 8 | use std::io::{Read, Seek}; [INFO] [stdout] | ^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused imports: `Result` and `Tokenizer` [INFO] [stdout] --> examples/transfomer_test.rs:16:29 [INFO] [stdout] | [INFO] [stdout] 16 | use tokenizers::tokenizer::{Result, Tokenizer}; [INFO] [stdout] | ^^^^^^ ^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `update_parameters` [INFO] [stdout] --> examples/transfomer_test.rs:398:12 [INFO] [stdout] | [INFO] [stdout] 398 | use poro::{update_parameters, Shape}; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `rand::Rng` [INFO] [stdout] --> examples/transfomer_test.rs:399:5 [INFO] [stdout] | [INFO] [stdout] 399 | use rand::Rng; [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unreachable statement [INFO] [stdout] --> examples/transfomer_test.rs:532:5 [INFO] [stdout] | [INFO] [stdout] 530 | panic!(""); [INFO] [stdout] | ---------- any code following this expression is unreachable [INFO] [stdout] 531 | [INFO] [stdout] 532 | let test_output = test_ouput.reshape(vec![batch_size * seq_length, 50257].into()); [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ unreachable statement [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unreachable_code)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `as_vec` [INFO] [stdout] --> examples/transfomer_test.rs:417:14 [INFO] [stdout] | [INFO] [stdout] 417 | let as_vec = read.read_to_end(&mut vec_buffer).unwrap(); [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_as_vec` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_variables)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `vocab_size` [INFO] [stdout] --> examples/transfomer_test.rs:508:9 [INFO] [stdout] | [INFO] [stdout] 508 | let vocab_size = 50304; [INFO] [stdout] | ^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_vocab_size` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `one_hot_encoded_trues` [INFO] [stdout] --> examples/transfomer_test.rs:525:9 [INFO] [stdout] | [INFO] [stdout] 525 | let one_hot_encoded_trues = one_hot_encoded_trues.reshape(vec![batch_size * seq_length, 50257].into()); [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_one_hot_encoded_trues` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `test_ouput` [INFO] [stdout] --> examples/transfomer_test.rs:528:9 [INFO] [stdout] | [INFO] [stdout] 528 | let test_ouput = gpt.forward(&x); [INFO] [stdout] | ^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_test_ouput` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `input` [INFO] [stdout] --> examples/transfomer_test.rs:565:9 [INFO] [stdout] | [INFO] [stdout] 565 | let input = Tensor::randn(vec![1, 1, 1].into()); [INFO] [stdout] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_input` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `output` [INFO] [stdout] --> examples/transfomer_test.rs:566:9 [INFO] [stdout] | [INFO] [stdout] 566 | let output = gpt.forward(&test_input); [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_output` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `gpt_config` [INFO] [stdout] --> examples/transfomer_test.rs:568:9 [INFO] [stdout] | [INFO] [stdout] 568 | let gpt_config = GPTConfig::from_bytestream(&mut std::fs::File::open("gpt2.bin").unwrap()); [INFO] [stdout] | ^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_gpt_config` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `position_encoder` [INFO] [stdout] --> examples/transfomer_test.rs:572:13 [INFO] [stdout] | [INFO] [stdout] 572 | let mut position_encoder = PositionalEncoding::new(10, 10); [INFO] [stdout] | ^^^^^^^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_position_encoder` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `embeddings` [INFO] [stdout] --> examples/transfomer_test.rs:587:13 [INFO] [stdout] | [INFO] [stdout] 587 | let mut embeddings = Embedding::from_tensor(embeddings_weight); [INFO] [stdout] | ^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_embeddings` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `write_f32_vector_to_file` is never used [INFO] [stdout] --> examples/transfomer_test.rs:17:4 [INFO] [stdout] | [INFO] [stdout] 17 | fn write_f32_vector_to_file(path: &str, data: &[f32]) -> std::io::Result<()> { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `write_string_vector_to_file` is never used [INFO] [stdout] --> examples/transfomer_test.rs:39:4 [INFO] [stdout] | [INFO] [stdout] 39 | fn write_string_vector_to_file(path: &str, data: &str) -> std::io::Result<()> { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: fields `block_size`, `vocab_size`, `number_of_heads`, and `embedding_dim` are never read [INFO] [stdout] --> examples/transfomer_test.rs:130:5 [INFO] [stdout] | [INFO] [stdout] 129 | struct GPTConfig { [INFO] [stdout] | --------- fields in this struct [INFO] [stdout] 130 | block_size: usize, [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] 131 | vocab_size: usize, [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] 132 | number_of_layers: usize, [INFO] [stdout] 133 | number_of_heads: usize, [INFO] [stdout] | ^^^^^^^^^^^^^^^ [INFO] [stdout] 134 | embedding_dim: usize [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `get_learning_rate` is never used [INFO] [stdout] --> examples/transfomer_test.rs:477:4 [INFO] [stdout] | [INFO] [stdout] 477 | fn get_learning_rate(iteration: usize) -> f32 { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unreachable statement [INFO] [stdout] --> examples/3_modules_and_models.rs:40:5 [INFO] [stdout] | [INFO] [stdout] 38 | return; [INFO] [stdout] | ------ any code following this expression is unreachable [INFO] [stdout] 39 | [INFO] [stdout] 40 | let linear_layer_config = LinearLayerConfig::new(3, 1); [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ unreachable statement [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unreachable_code)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variable does not need to be mutable [INFO] [stdout] --> examples/3_modules_and_models.rs:23:9 [INFO] [stdout] | [INFO] [stdout] 23 | let mut tensor_from = Tensor::from_vec(vec![1.0, 2.0, 3.0, 4.0], vec![2, 2].into()); [INFO] [stdout] | ----^^^^^^^^^^^ [INFO] [stdout] | | [INFO] [stdout] | help: remove this `mut` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_mut)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variable does not need to be mutable [INFO] [stdout] --> examples/3_modules_and_models.rs:24:9 [INFO] [stdout] | [INFO] [stdout] 24 | let mut bias = Tensor::from_vec(vec![0.0, 0.0], vec![2].into()); [INFO] [stdout] | ----^^^^ [INFO] [stdout] | | [INFO] [stdout] | help: remove this `mut` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unreachable statement [INFO] [stdout] --> src/nn/layers/embedding.rs:31:9 [INFO] [stdout] | [INFO] [stdout] 29 | return self.tensor.embbeding(input, self.model_dimension); [INFO] [stdout] | --------------------------------------------------------- any code following this expression is unreachable [INFO] [stdout] 30 | [INFO] [stdout] 31 | let mut test_index_tensor = Tensor::zeroes(Shape::new(vec![input.shape.indices[0], input.shape.indices[1], self.model_dimension])); [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ unreachable statement [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unreachable_code)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unreachable statement [INFO] [stdout] --> src/nn/layers/layer_norm.rs:200:9 [INFO] [stdout] | [INFO] [stdout] 198 | ... panic!("---"); [INFO] [stdout] | ------------- any code following this expression is unreachable [INFO] [stdout] 199 | ... [INFO] [stdout] 200 | ... let expected_layer_norm_weight_grad_output_flatten = expected_layer_norm_weight_grad.item().iter().map(|x| x.clone()).collect::>(); [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ unreachable statement [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `test_input` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:157:13 [INFO] [stdout] | [INFO] [stdout] 157 | let test_input = Tensor::from_bytestream(&mut test_input_file, false).unwrap(); [INFO] [stdout] | ^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_test_input` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_variables)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `expected_layer_norm_weight_grad` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:163:13 [INFO] [stdout] | [INFO] [stdout] 163 | let expected_layer_norm_weight_grad = Tensor::from_bytestream(&mut layer_norm_weight_grad_file, false).unwrap(); [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_expected_layer_norm_weight_grad` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `layer_norm_bias_grad` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:164:13 [INFO] [stdout] | [INFO] [stdout] 164 | let layer_norm_bias_grad = Tensor::from_bytestream(&mut layer_norm_bias_grad_file, false).unwrap(); [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_layer_norm_bias_grad` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `o` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:179:14 [INFO] [stdout] | [INFO] [stdout] 179 | for (o, e) in ouput_as_flat_array.iter().zip(expected_as_flat_array.iter()) { [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_o` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `e` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:179:17 [INFO] [stdout] | [INFO] [stdout] 179 | for (o, e) in ouput_as_flat_array.iter().zip(expected_as_flat_array.iter()) { [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_e` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `loss` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:189:14 [INFO] [stdout] | [INFO] [stdout] 189 | for (loss, expected) in mse_loss.item().iter().zip(expected_mse_loss.iter()) { [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_loss` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `expected` [INFO] [stdout] --> src/nn/layers/layer_norm.rs:189:20 [INFO] [stdout] | [INFO] [stdout] 189 | for (loss, expected) in mse_loss.item().iter().zip(expected_mse_loss.iter()) { [INFO] [stdout] | ^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_expected` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `linear_layer` [INFO] [stdout] --> src/nn/layers/linear.rs:63:17 [INFO] [stdout] | [INFO] [stdout] 63 | let mut linear_layer = LinearLayer::new(config); [INFO] [stdout] | ^^^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_linear_layer` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variable does not need to be mutable [INFO] [stdout] --> src/nn/layers/linear.rs:63:13 [INFO] [stdout] | [INFO] [stdout] 63 | let mut linear_layer = LinearLayer::new(config); [INFO] [stdout] | ----^^^^^^^^^^^^ [INFO] [stdout] | | [INFO] [stdout] | help: remove this `mut` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_mut)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `source_grad` [INFO] [stdout] --> src/central/equation.rs:805:21 [INFO] [stdout] | [INFO] [stdout] 805 | let source_grad = self.get_tensor_grad(origin_tensor).into_raw_vec(); [INFO] [stdout] | ^^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_source_grad` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `vocab_size` is never read [INFO] [stdout] --> src/nn/layers/embedding.rs:7:5 [INFO] [stdout] | [INFO] [stdout] 5 | pub struct Embedding { [INFO] [stdout] | --------- field in this struct [INFO] [stdout] 6 | pub tensor: Tensor, [INFO] [stdout] 7 | vocab_size: usize, [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: struct `MLPConfig` is never constructed [INFO] [stdout] --> src/nn/layers/mlp.rs:23:8 [INFO] [stdout] | [INFO] [stdout] 23 | struct MLPConfig { [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variable `M` should have a snake case name [INFO] [stdout] --> src/central/equation.rs:431:9 [INFO] [stdout] | [INFO] [stdout] 431 | let M = a_shape[a_shape.len() - 2]; [INFO] [stdout] | ^ help: convert the identifier to snake case: `m` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(non_snake_case)]` (part of `#[warn(nonstandard_style)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variable `K_a` should have a snake case name [INFO] [stdout] --> src/central/equation.rs:432:9 [INFO] [stdout] | [INFO] [stdout] 432 | let K_a = a_shape[a_shape.len() - 1]; [INFO] [stdout] | ^^^ help: convert the identifier to snake case (notice the capitalization): `k_a` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: associated function `new` is never used [INFO] [stdout] --> src/nn/layers/mlp.rs:34:8 [INFO] [stdout] | [INFO] [stdout] 33 | impl MLP { [INFO] [stdout] | -------- associated function in this implementation [INFO] [stdout] 34 | fn new(config: MLPConfig) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `length_of_normalized_shape` is never read [INFO] [stdout] --> src/nn/layers/layer_norm.rs:9:5 [INFO] [stdout] | [INFO] [stdout] 6 | pub struct LayerNorm { [INFO] [stdout] | --------- field in this struct [INFO] [stdout] ... [INFO] [stdout] 9 | length_of_normalized_shape: usize, [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variable `K_b` should have a snake case name [INFO] [stdout] --> src/central/equation.rs:433:9 [INFO] [stdout] | [INFO] [stdout] 433 | let K_b = b_shape[b_shape.len() - 2]; [INFO] [stdout] | ^^^ help: convert the identifier to snake case (notice the capitalization): `k_b` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variable `N` should have a snake case name [INFO] [stdout] --> src/central/equation.rs:434:9 [INFO] [stdout] | [INFO] [stdout] 434 | let N = b_shape[b_shape.len() - 1]; [INFO] [stdout] | ^ help: convert the identifier to snake case: `n` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: method `add_time` is never used [INFO] [stdout] --> src/central/equation.rs:201:8 [INFO] [stdout] | [INFO] [stdout] 46 | impl Equation { [INFO] [stdout] | ------------- method in this implementation [INFO] [stdout] ... [INFO] [stdout] 201 | fn add_time(&mut self, operation: &str, now: Instant) { [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `write_f32_vector_to_file` is never used [INFO] [stdout] --> src/nn/layers/casual_self_attention.rs:7:4 [INFO] [stdout] | [INFO] [stdout] 7 | fn write_f32_vector_to_file(path: &str, data: &[f32]) -> std::io::Result<()> { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `write_string_vector_to_file` is never used [INFO] [stdout] --> src/nn/layers/casual_self_attention.rs:29:4 [INFO] [stdout] | [INFO] [stdout] 29 | fn write_string_vector_to_file(path: &str, data: &str) -> std::io::Result<()> { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: associated function `transpose_inside` is never used [INFO] [stdout] --> src/central/tensor.rs:632:8 [INFO] [stdout] | [INFO] [stdout] 45 | impl Tensor { [INFO] [stdout] | ----------- associated function in this implementation [INFO] [stdout] ... [INFO] [stdout] 632 | fn transpose_inside(matrix: Vec, rows: usize, cols: usize) -> Vec { [INFO] [stdout] | ^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] Finished `test` profile [unoptimized + debuginfo] target(s) in 6.55s [INFO] running `Command { std: "docker" "inspect" "7edaba70dea76f816c8a4506f2d5e176c9b84f2c72c1fdba07bbe12cb7796183", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "7edaba70dea76f816c8a4506f2d5e176c9b84f2c72c1fdba07bbe12cb7796183", kill_on_drop: false }` [INFO] [stdout] 7edaba70dea76f816c8a4506f2d5e176c9b84f2c72c1fdba07bbe12cb7796183