[INFO] cloning repository https://github.com/comfortablynumb/myllm
[INFO] running `Command { std: "git" "-c" "credential.helper=" "-c" "credential.helper=/workspace/cargo-home/bin/git-credential-null" "clone" "--bare" "https://github.com/comfortablynumb/myllm" "/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2Fcomfortablynumb%2Fmyllm", kill_on_drop: false }`
[INFO] [stderr] Cloning into bare repository '/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2Fcomfortablynumb%2Fmyllm'...
[INFO] running `Command { std: "git" "rev-parse" "HEAD", kill_on_drop: false }`
[INFO] [stdout] 3f4031598d5eba236544b8bfcd136cf3c4163221
[INFO] testing comfortablynumb/myllm against 1.95.0 for beta-1.96-2
[INFO] running `Command { std: "git" "clone" "/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2Fcomfortablynumb%2Fmyllm" "/workspace/builds/worker-0-tc1/source", kill_on_drop: false }`
[INFO] [stderr] Cloning into '/workspace/builds/worker-0-tc1/source'...
[INFO] [stderr] done.
[INFO] started tweaking git repo https://github.com/comfortablynumb/myllm
[INFO] finished tweaking git repo https://github.com/comfortablynumb/myllm
[INFO] tweaked toml for git repo https://github.com/comfortablynumb/myllm written to /workspace/builds/worker-0-tc1/source/Cargo.toml
[INFO] validating manifest of git repo https://github.com/comfortablynumb/myllm on toolchain 1.95.0
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+1.95.0" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }`
[INFO] crate git repo https://github.com/comfortablynumb/myllm already has a lockfile, it will not be regenerated
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+1.95.0" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }`
[INFO] [stderr]     Updating crates.io index
[INFO] [stderr]  Downloading crates ...
[INFO] [stderr]   Downloaded safetensors v0.4.5
[INFO] [stderr]   Downloaded smallvec v2.0.0-alpha.10
[INFO] [stderr]   Downloaded ort-sys v2.0.0-rc.10
[INFO] [stderr]   Downloaded openvino v0.7.3
[INFO] [stderr]   Downloaded openvino-finder v0.7.3
[INFO] [stderr]   Downloaded openvino-sys v0.7.3
[INFO] [stderr]   Downloaded zmij v0.1.9
[INFO] [stderr]   Downloaded metrics-exporter-prometheus v0.16.2
[INFO] [stderr]   Downloaded axum-extra v0.10.3
[INFO] [stderr]   Downloaded axum-test v17.3.0
[INFO] [stderr]   Downloaded ort v2.0.0-rc.10
[INFO] [stderr]   Downloaded serde_json v1.0.147
[INFO] [stderr]   Downloaded metrics-util v0.19.1
[INFO] [stderr]   Downloaded aws-lc-rs v1.15.2
[INFO] [stderr]   Downloaded cudarc v0.12.1
[INFO] [stderr]   Downloaded aws-lc-sys v0.35.0
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-0-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-0-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:d429b63d4308055ea97f60fb1d3dfca48854a00942f1bd2ad806beaf015945ec" "/opt/rustwide/cargo-home/bin/cargo" "+1.95.0" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }`
[INFO] [stdout] 6ac4a02d8961e4a9e531d44bc98e26925a2e2d956dbeb4b7a6b6c83dae1b2a16
[INFO] running `Command { std: "docker" "start" "-a" "6ac4a02d8961e4a9e531d44bc98e26925a2e2d956dbeb4b7a6b6c83dae1b2a16", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "inspect" "6ac4a02d8961e4a9e531d44bc98e26925a2e2d956dbeb4b7a6b6c83dae1b2a16", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "6ac4a02d8961e4a9e531d44bc98e26925a2e2d956dbeb4b7a6b6c83dae1b2a16", kill_on_drop: false }`
[INFO] [stdout] 6ac4a02d8961e4a9e531d44bc98e26925a2e2d956dbeb4b7a6b6c83dae1b2a16
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-0-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-0-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=warn" "-e" "RUSTDOCFLAGS=--cap-lints=warn" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:d429b63d4308055ea97f60fb1d3dfca48854a00942f1bd2ad806beaf015945ec" "/opt/rustwide/cargo-home/bin/cargo" "+1.95.0" "build" "--frozen" "--message-format=json", kill_on_drop: false }`
[INFO] [stdout] 3b8a792b6f56861e5eeeb785ed1a4504b056c622947088d978a17140fc70698a
[INFO] running `Command { std: "docker" "start" "-a" "3b8a792b6f56861e5eeeb785ed1a4504b056c622947088d978a17140fc70698a", kill_on_drop: false }`
[INFO] [stderr]    Compiling proc-macro2 v1.0.103
[INFO] [stderr]    Compiling unicode-ident v1.0.22
[INFO] [stderr]    Compiling quote v1.0.42
[INFO] [stderr]    Compiling libc v0.2.178
[INFO] [stderr]    Compiling cfg-if v1.0.4
[INFO] [stderr]    Compiling autocfg v1.5.0
[INFO] [stderr]    Compiling serde_core v1.0.228
[INFO] [stderr]    Compiling serde v1.0.228
[INFO] [stderr]    Compiling memchr v2.7.6
[INFO] [stderr]    Compiling crossbeam-utils v0.8.21
[INFO] [stderr]    Compiling libm v0.2.15
[INFO] [stderr]    Compiling zerocopy v0.8.31
[INFO] [stderr]    Compiling num-traits v0.2.19
[INFO] [stderr]    Compiling rayon-core v1.13.0
[INFO] [stderr]    Compiling crossbeam-epoch v0.9.18
[INFO] [stderr]    Compiling once_cell v1.21.3
[INFO] [stderr]    Compiling crossbeam-deque v0.8.6
[INFO] [stderr]    Compiling matrixmultiply v0.3.10
[INFO] [stderr]    Compiling thiserror v1.0.69
[INFO] [stderr]    Compiling pin-project-lite v0.2.16
[INFO] [stderr]    Compiling syn v2.0.111
[INFO] [stderr]    Compiling either v1.15.0
[INFO] [stderr]    Compiling getrandom v0.2.16
[INFO] [stderr]    Compiling rand_core v0.6.4
[INFO] [stderr]    Compiling aho-corasick v1.1.4
[INFO] [stderr]    Compiling rawpointer v0.2.1
[INFO] [stderr]    Compiling rayon v1.11.0
[INFO] [stderr]    Compiling parking_lot_core v0.9.12
[INFO] [stderr]    Compiling regex-syntax v0.8.8
[INFO] [stderr]    Compiling lazy_static v1.5.0
[INFO] [stderr]    Compiling smallvec v1.15.1
[INFO] [stderr]    Compiling ppv-lite86 v0.2.21
[INFO] [stderr]    Compiling num-integer v0.1.46
[INFO] [stderr]    Compiling num-complex v0.4.6
[INFO] [stderr]    Compiling rand_chacha v0.3.1
[INFO] [stderr]    Compiling scopeguard v1.2.0
[INFO] [stderr]    Compiling serde_json v1.0.147
[INFO] [stderr]    Compiling rand v0.8.5
[INFO] [stderr]    Compiling portable-atomic v1.12.0
[INFO] [stderr]    Compiling lock_api v0.4.14
[INFO] [stderr]    Compiling tracing-core v0.1.36
[INFO] [stderr]    Compiling futures-core v0.3.31
[INFO] [stderr]    Compiling zmij v0.1.9
[INFO] [stderr]    Compiling itoa v1.0.16
[INFO] [stderr]    Compiling futures-sink v0.3.31
[INFO] [stderr]    Compiling rand_distr v0.4.3
[INFO] [stderr]    Compiling unicode-width v0.2.2
[INFO] [stderr]    Compiling regex-automata v0.4.13
[INFO] [stderr]    Compiling ndarray v0.16.1
[INFO] [stderr]    Compiling console v0.15.11
[INFO] [stderr]    Compiling futures-channel v0.3.31
[INFO] [stderr]    Compiling parking_lot v0.12.5
[INFO] [stderr]    Compiling futures-task v0.3.31
[INFO] [stderr]    Compiling utf8parse v0.2.2
[INFO] [stderr]    Compiling equivalent v1.0.2
[INFO] [stderr]    Compiling pin-utils v0.1.0
[INFO] [stderr]    Compiling hashbrown v0.16.1
[INFO] [stderr]    Compiling futures-io v0.3.31
[INFO] [stderr]    Compiling slab v0.4.11
[INFO] [stderr]    Compiling number_prefix v0.4.0
[INFO] [stderr]    Compiling anstyle-parse v0.2.7
[INFO] [stderr]    Compiling indicatif v0.17.11
[INFO] [stderr]    Compiling errno v0.3.14
[INFO] [stderr]    Compiling anstyle-query v1.1.5
[INFO] [stderr]    Compiling anstyle v1.0.13
[INFO] [stderr]    Compiling colorchoice v1.0.4
[INFO] [stderr]    Compiling is_terminal_polyfill v1.70.2
[INFO] [stderr]    Compiling toml_write v0.1.2
[INFO] [stderr]    Compiling rustc-hash v2.1.1
[INFO] [stderr]    Compiling winnow v0.7.14
[INFO] [stderr]    Compiling anyhow v1.0.100
[INFO] [stderr]    Compiling anstream v0.6.21
[INFO] [stderr]    Compiling indexmap v2.12.1
[INFO] [stderr]    Compiling signal-hook-registry v1.4.8
[INFO] [stderr]    Compiling socket2 v0.6.1
[INFO] [stderr]    Compiling mio v1.1.1
[INFO] [stderr]    Compiling clap_lex v0.7.6
[INFO] [stderr]    Compiling serde_derive v1.0.228
[INFO] [stderr]    Compiling thiserror-impl v1.0.69
[INFO] [stderr]    Compiling futures-macro v0.3.31
[INFO] [stderr]    Compiling tracing-attributes v0.1.31
[INFO] [stderr]    Compiling tokio-macros v2.6.0
[INFO] [stderr]    Compiling bytes v1.11.0
[INFO] [stderr]    Compiling futures-util v0.3.31
[INFO] [stderr]    Compiling log v0.4.29
[INFO] [stderr]    Compiling tracing v0.1.44
[INFO] [stderr]    Compiling regex v1.12.2
[INFO] [stderr]    Compiling strsim v0.11.1
[INFO] [stderr]    Compiling heck v0.5.0
[INFO] [stderr]    Compiling clap_derive v4.5.49
[INFO] [stderr]    Compiling clap_builder v4.5.53
[INFO] [stderr]    Compiling tokio v1.48.0
[INFO] [stderr]    Compiling tracing-log v0.2.0
[INFO] [stderr]    Compiling matchers v0.2.0
[INFO] [stderr]    Compiling sharded-slab v0.1.7
[INFO] [stderr]    Compiling memmap2 v0.9.9
[INFO] [stderr]    Compiling thread_local v1.1.9
[INFO] [stderr]    Compiling nu-ansi-term v0.50.3
[INFO] [stderr]    Compiling tracing-subscriber v0.3.22
[INFO] [stderr]    Compiling clap v4.5.53
[INFO] [stderr]    Compiling futures-executor v0.3.31
[INFO] [stderr]    Compiling futures v0.3.31
[INFO] [stderr]    Compiling myllm-core v0.1.0 (/opt/rustwide/workdir/crates/myllm-core)
[INFO] [stderr]    Compiling toml_datetime v0.6.11
[INFO] [stderr]    Compiling serde_spanned v0.6.9
[INFO] [stderr]    Compiling safetensors v0.4.5
[INFO] [stderr]    Compiling myllm-tokenizer v0.1.0 (/opt/rustwide/workdir/crates/myllm-tokenizer)
[INFO] [stderr]    Compiling toml_edit v0.22.27
[INFO] [stdout] warning: field `config` is never read
[INFO] [stdout]    --> crates/myllm-core/src/quantization/calibration.rs:283:5
[INFO] [stdout]     |
[INFO] [stdout] 280 | pub struct ModelCalibrator {
[INFO] [stdout]     |            --------------- field in this struct
[INFO] [stdout] ...
[INFO] [stdout] 283 |     config: CalibratorConfig,
[INFO] [stdout]     |     ^^^^^^
[INFO] [stdout]     |
[INFO] [stdout]     = note: `ModelCalibrator` has a derived impl for the trait `Debug`, but this is intentionally ignored during dead code analysis
[INFO] [stdout]     = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: fields `max_abs` and `errors_by_bits` are never read
[INFO] [stdout]    --> crates/myllm-core/src/quantization/exl2.rs:177:5
[INFO] [stdout]     |
[INFO] [stdout] 171 | struct GroupStats {
[INFO] [stdout]     |        ---------- fields in this struct
[INFO] [stdout] ...
[INFO] [stdout] 177 |     max_abs: f32,
[INFO] [stdout]     |     ^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout] 181 |     errors_by_bits: Vec<f32>,
[INFO] [stdout]     |     ^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout]     = note: `GroupStats` has derived impls for the traits `Clone` and `Debug`, but these are intentionally ignored during dead code analysis
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stderr]    Compiling myllm-nn v0.1.0 (/opt/rustwide/workdir/crates/myllm-nn)
[INFO] [stdout] warning: unused variable: `fan_in`
[INFO] [stdout]   --> crates/myllm-nn/src/conv.rs:99:13
[INFO] [stdout]    |
[INFO] [stdout] 99 |         let fan_in = config.in_channels * kh * kw / config.groups;
[INFO] [stdout]    |             ^^^^^^ help: if this is intentional, prefix it with an underscore: `_fan_in`
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(unused_variables)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `batch`
[INFO] [stdout]    --> crates/myllm-nn/src/conv.rs:155:13
[INFO] [stdout]     |
[INFO] [stdout] 155 |         let batch = shape.dim(0);
[INFO] [stdout]     |             ^^^^^ help: if this is intentional, prefix it with an underscore: `_batch`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `sh`
[INFO] [stdout]    --> crates/myllm-nn/src/conv.rs:161:14
[INFO] [stdout]     |
[INFO] [stdout] 161 |         let (sh, sw) = self.config.stride;
[INFO] [stdout]     |              ^^ help: if this is intentional, prefix it with an underscore: `_sh`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `sw`
[INFO] [stdout]    --> crates/myllm-nn/src/conv.rs:161:18
[INFO] [stdout]     |
[INFO] [stdout] 161 |         let (sh, sw) = self.config.stride;
[INFO] [stdout]     |                  ^^ help: if this is intentional, prefix it with an underscore: `_sw`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `col_size`
[INFO] [stdout]    --> crates/myllm-nn/src/conv.rs:176:13
[INFO] [stdout]     |
[INFO] [stdout] 176 |         let col_size = channels_per_group * kh * kw;
[INFO] [stdout]     |             ^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_col_size`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused variable: `spatial_size`
[INFO] [stdout]    --> crates/myllm-nn/src/conv.rs:177:13
[INFO] [stdout]     |
[INFO] [stdout] 177 |         let spatial_size = out_h * out_w;
[INFO] [stdout]     |             ^^^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_spatial_size`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: method `unfold` is never used
[INFO] [stdout]    --> crates/myllm-nn/src/conv.rs:153:8
[INFO] [stdout]     |
[INFO] [stdout]  95 | impl<B: Backend> Conv2d<B> {
[INFO] [stdout]     | -------------------------- method in this implementation
[INFO] [stdout] ...
[INFO] [stdout] 153 |     fn unfold(&self, input: &Tensor<B>) -> Result<Tensor<B>> {
[INFO] [stdout]     |        ^^^^^^
[INFO] [stdout]     |
[INFO] [stdout]     = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: field `direction_base` is never read
[INFO] [stdout]   --> crates/myllm-nn/src/dora.rs:87:5
[INFO] [stdout]    |
[INFO] [stdout] 83 | pub struct DoraLinear<B: Backend = Cpu> {
[INFO] [stdout]    |            ---------- field in this struct
[INFO] [stdout] ...
[INFO] [stdout] 87 |     direction_base: Tensor<B>,
[INFO] [stdout]    |     ^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: field `dropout` is never read
[INFO] [stdout]    --> crates/myllm-nn/src/lora.rs:123:5
[INFO] [stdout]     |
[INFO] [stdout] 111 | pub struct LoraLinear<B: Backend = Cpu> {
[INFO] [stdout]     |            ---------- field in this struct
[INFO] [stdout] ...
[INFO] [stdout] 123 |     dropout: f32,
[INFO] [stdout]     |     ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: field `config` is never read
[INFO] [stdout]    --> crates/myllm-nn/src/qlora.rs:354:5
[INFO] [stdout]     |
[INFO] [stdout] 350 | pub struct QLoraAdapter<B: Backend = Cpu> {
[INFO] [stdout]     |            ------------ field in this struct
[INFO] [stdout] ...
[INFO] [stdout] 354 |     config: QLoraConfig,
[INFO] [stdout]     |     ^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stderr]    Compiling myllm-model v0.1.0 (/opt/rustwide/workdir/crates/myllm-model)
[INFO] [stdout] error[E0433]: cannot find `CoreError` in `myllm_core`
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:729:25
[INFO] [stdout]     |
[INFO] [stdout] 729 |             myllm_core::CoreError::InvalidShape("No pooler output available".to_string())
[INFO] [stdout]     |                         ^^^^^^^^^ could not find `CoreError` in `myllm_core`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0433]: cannot find `CoreError` in `myllm_core`
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:504:25
[INFO] [stdout]     |
[INFO] [stdout] 504 |             myllm_core::CoreError::InvalidShape("No pooler output".to_string())
[INFO] [stdout]     |                         ^^^^^^^^^ could not find `CoreError` in `myllm_core`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `RMSNorm`
[INFO] [stdout]   --> crates/myllm-model/src/encoder.rs:10:76
[INFO] [stdout]    |
[INFO] [stdout] 10 | use myllm_nn::{Dropout, Embedding, Layer, LayerNorm, Linear, LinearConfig, RMSNorm};
[INFO] [stdout]    |                                                                            ^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `LayerNorm`
[INFO] [stdout]   --> crates/myllm-model/src/encoder_decoder.rs:10:43
[INFO] [stdout]    |
[INFO] [stdout] 10 | use myllm_nn::{Dropout, Embedding, Layer, LayerNorm, Linear, LinearConfig, RMSNorm};
[INFO] [stdout]    |                                           ^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `crate::rope::RotaryEmbedding`
[INFO] [stdout]   --> crates/myllm-model/src/encoder_decoder.rs:15:5
[INFO] [stdout]    |
[INFO] [stdout] 15 | use crate::rope::RotaryEmbedding;
[INFO] [stdout]    |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `unwrap` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:163:56
[INFO] [stdout]     |
[INFO] [stdout] 163 |             Tensor::randn([out_features, in_features]).unwrap() * scale;
[INFO] [stdout]     |                                                        ^^^^^^ method not found in `Tensor<_>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `unwrap` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:166:48
[INFO] [stdout]     |
[INFO] [stdout] 166 |             Some(Tensor::zeros([out_features]).unwrap())
[INFO] [stdout]     |                                                ^^^^^^ method not found in `Tensor<_>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try`
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:196:21
[INFO] [stdout]     |
[INFO] [stdout] 196 |         let zeros = Tensor::zeros(tensor.shape().dims())?;
[INFO] [stdout]     |                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ the `?` operator cannot be applied to type `Tensor<_>`
[INFO] [stdout]     |
[INFO] [stdout]     = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor<_>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try`
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:197:20
[INFO] [stdout]     |
[INFO] [stdout] 197 |         let ones = Tensor::ones(tensor.shape().dims())?;
[INFO] [stdout]     |                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ the `?` operator cannot be applied to type `Tensor<_>`
[INFO] [stdout]     |
[INFO] [stdout]     = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor<_>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: the method `gt` exists for reference `&Tensor<B>`, but its trait bounds were not satisfied
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:200:31
[INFO] [stdout]     |
[INFO] [stdout] 200 |         let positive = tensor.gt(&zeros)?;
[INFO] [stdout]     |                               ^^ method cannot be called on `&Tensor<B>` due to unsatisfied trait bounds
[INFO] [stdout]     |
[INFO] [stdout]    ::: crates/myllm-core/src/tensor.rs:12:1
[INFO] [stdout]     |
[INFO] [stdout]  12 | pub struct Tensor<B: Backend = Cpu> {
[INFO] [stdout]     | ----------------------------------- doesn't satisfy `Tensor<B>: Iterator` or `Tensor<B>: PartialOrd<_>`
[INFO] [stdout]     |
[INFO] [stdout]     = note: the following trait bounds were not satisfied:
[INFO] [stdout]             `Tensor<B>: PartialOrd<_>`
[INFO] [stdout]             which is required by `&Tensor<B>: PartialOrd<&_>`
[INFO] [stdout]             `&Tensor<B>: Iterator`
[INFO] [stdout]             which is required by `&mut &Tensor<B>: Iterator`
[INFO] [stdout]             `Tensor<B>: Iterator`
[INFO] [stdout]             which is required by `&mut Tensor<B>: Iterator`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `div_scalar` found for reference `&Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:206:29
[INFO] [stdout]     |
[INFO] [stdout] 206 |         let scaled = tensor.div_scalar(scale)?;
[INFO] [stdout]     |                             ^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: there is a method `add_scalar` with a similar name
[INFO] [stdout]     |
[INFO] [stdout] 206 -         let scaled = tensor.div_scalar(scale)?;
[INFO] [stdout] 206 +         let scaled = tensor.add_scalar(scale)?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `abs` found for reference `&Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:216:26
[INFO] [stdout]     |
[INFO] [stdout] 216 |         let abs = tensor.abs()?;
[INFO] [stdout]     |                          ^^^ method not found in `&Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `div_scalar` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:232:46
[INFO] [stdout]     |
[INFO] [stdout] 232 |                 let normalized = self.weight.div_scalar(scale)?;
[INFO] [stdout]     |                                              ^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: there is a method `add_scalar` with a similar name
[INFO] [stdout]     |
[INFO] [stdout] 232 -                 let normalized = self.weight.div_scalar(scale)?;
[INFO] [stdout] 232 +                 let normalized = self.weight.add_scalar(scale)?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `abs` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:237:46
[INFO] [stdout]     |
[INFO] [stdout] 237 |                 let abs_weight = self.weight.abs()?;
[INFO] [stdout]     |                                              ^^^ method not found in `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `div_scalar` found for reference `&Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:260:40
[INFO] [stdout]     |
[INFO] [stdout] 260 |                 let normalized = input.div_scalar(scale)?;
[INFO] [stdout]     |                                        ^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: there is a method `add_scalar` with a similar name
[INFO] [stdout]     |
[INFO] [stdout] 260 -                 let normalized = input.div_scalar(scale)?;
[INFO] [stdout] 260 +                 let normalized = input.add_scalar(scale)?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this method takes 0 arguments but 2 arguments were supplied
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:273:53
[INFO] [stdout]     |
[INFO] [stdout] 273 |         let output = input.matmul(&quantized_weight.transpose(0, 1)?)?;
[INFO] [stdout]     |                                                     ^^^^^^^^^ -  - unexpected argument #2 of type `{integer}`
[INFO] [stdout]     |                                                               |
[INFO] [stdout]     |                                                               unexpected argument #1 of type `{integer}`
[INFO] [stdout]     |
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:149:12
[INFO] [stdout]     |
[INFO] [stdout] 149 |     pub fn transpose(&self) -> Result<Self> {
[INFO] [stdout]     |            ^^^^^^^^^
[INFO] [stdout] help: remove the extra arguments
[INFO] [stdout]     |
[INFO] [stdout] 273 -         let output = input.matmul(&quantized_weight.transpose(0, 1)?)?;
[INFO] [stdout] 273 +         let output = input.matmul(&quantized_weight.transpose()?)?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `unwrap` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:329:49
[INFO] [stdout]     |
[INFO] [stdout] 329 |             weight: Tensor::ones([hidden_size]).unwrap(),
[INFO] [stdout]     |                                                 ^^^^^^ method not found in `Tensor<_>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `pow_scalar` found for reference `&Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:336:26
[INFO] [stdout]     |
[INFO] [stdout] 336 |         let variance = x.pow_scalar(2.0)?.mean_dim(-1, true)?;
[INFO] [stdout]     |                          ^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: there is a method `add_scalar` with a similar name
[INFO] [stdout]     |
[INFO] [stdout] 336 -         let variance = x.pow_scalar(2.0)?.mean_dim(-1, true)?;
[INFO] [stdout] 336 +         let variance = x.add_scalar(2.0)?.mean_dim(-1, true)?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `mul` found for reference `&Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:339:28
[INFO] [stdout]     |
[INFO] [stdout] 339 |         let normalized = x.mul(&rsqrt)?;
[INFO] [stdout]     |                            ^^^ this is an associated function, not a method
[INFO] [stdout]     |
[INFO] [stdout]     = note: found the following associated functions; to be used as methods, functions must have a `self` parameter
[INFO] [stdout] note: the candidate is defined in the trait `Backend`
[INFO] [stdout]    --> crates/myllm-core/src/backend/mod.rs:64:5
[INFO] [stdout]     |
[INFO] [stdout]  64 |     fn mul(a: &Self::Storage, b: &Self::Storage) -> Result<Self::Storage>;
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     = help: items from traits can only be used if the trait is in scope
[INFO] [stdout] help: use associated function syntax instead
[INFO] [stdout]     |
[INFO] [stdout] 339 -         let normalized = x.mul(&rsqrt)?;
[INFO] [stdout] 339 +         let normalized = Tensor<B>::mul(&rsqrt)?;
[INFO] [stdout]     |
[INFO] [stdout] help: trait `Mul` which provides `mul` is implemented but not in scope; perhaps you want to import it
[INFO] [stdout]     |
[INFO] [stdout]  12 + use std::ops::Mul;
[INFO] [stdout]     |
[INFO] [stdout] help: there is a method `matmul` with a similar name
[INFO] [stdout]     |
[INFO] [stdout] 339 |         let normalized = x.matmul(&rsqrt)?;
[INFO] [stdout]     |                            +++
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `mul_scalar` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:429:25
[INFO] [stdout]     |
[INFO] [stdout] 429 |         scores = scores.mul_scalar(self.scale)?;
[INFO] [stdout]     |                         ^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: there is a method `add_scalar` with a similar name
[INFO] [stdout]     |
[INFO] [stdout] 429 -         scores = scores.mul_scalar(self.scale)?;
[INFO] [stdout] 429 +         scores = scores.add_scalar(self.scale)?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no function or associated item named `full` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:432:35
[INFO] [stdout]     |
[INFO] [stdout] 432 |             let neg_inf = Tensor::full(scores.shape().dims(), -1e9)?;
[INFO] [stdout]     |                                   ^^^^ function or associated item not found in `Tensor<_>`
[INFO] [stdout]     |
[INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions:
[INFO] [stdout]       Tensor::<B>::from_storage
[INFO] [stdout]       Tensor::<B>::from_slice
[INFO] [stdout]       Tensor::<B>::zeros
[INFO] [stdout]       Tensor::<B>::ones
[INFO] [stdout]       and 7 others
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:20:5
[INFO] [stdout]     |
[INFO] [stdout]  20 |     pub fn from_storage(storage: B::Storage) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  29 |     pub fn from_slice(data: &[f32], shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  35 |     pub fn zeros(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  40 |     pub fn ones(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `unsqueeze` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:436:43
[INFO] [stdout]     |
[INFO] [stdout] 436 |             let causal_mask = causal_mask.unsqueeze(0)?.unsqueeze(0)?;
[INFO] [stdout]     |                                           ^^^^^^^^^ method not found in `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no function or associated item named `full` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:437:35
[INFO] [stdout]     |
[INFO] [stdout] 437 |             let neg_inf = Tensor::full(scores.shape().dims(), -1e9)?;
[INFO] [stdout]     |                                   ^^^^ function or associated item not found in `Tensor<_>`
[INFO] [stdout]     |
[INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions:
[INFO] [stdout]       Tensor::<B>::from_storage
[INFO] [stdout]       Tensor::<B>::from_slice
[INFO] [stdout]       Tensor::<B>::zeros
[INFO] [stdout]       Tensor::<B>::ones
[INFO] [stdout]       and 7 others
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:20:5
[INFO] [stdout]     |
[INFO] [stdout]  20 |     pub fn from_storage(storage: B::Storage) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  29 |     pub fn from_slice(data: &[f32], shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  35 |     pub fn zeros(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  40 |     pub fn ones(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this method takes 0 arguments but 1 argument was supplied
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:441:35
[INFO] [stdout]     |
[INFO] [stdout] 441 |         let attn_weights = scores.softmax(-1)?;
[INFO] [stdout]     |                                   ^^^^^^^ -- unexpected argument of type `{integer}`
[INFO] [stdout]     |
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:189:12
[INFO] [stdout]     |
[INFO] [stdout] 189 |     pub fn softmax(&self) -> Self {
[INFO] [stdout]     |            ^^^^^^^
[INFO] [stdout] help: remove the extra argument
[INFO] [stdout]     |
[INFO] [stdout] 441 -         let attn_weights = scores.softmax(-1)?;
[INFO] [stdout] 441 +         let attn_weights = scores.softmax()?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try`
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:441:28
[INFO] [stdout]     |
[INFO] [stdout] 441 |         let attn_weights = scores.softmax(-1)?;
[INFO] [stdout]     |                            ^^^^^^^^^^^^^^^^^^^ the `?` operator cannot be applied to type `Tensor<B>`
[INFO] [stdout]     |
[INFO] [stdout]     = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try`
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:511:34
[INFO] [stdout]     |
[INFO] [stdout] 511 |             let gate_activated = gate.silu()?;
[INFO] [stdout]     |                                  ^^^^^^^^^^^^ the `?` operator cannot be applied to type `Tensor<B>`
[INFO] [stdout]     |
[INFO] [stdout]     = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try`
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:520:13
[INFO] [stdout]     |
[INFO] [stdout] 520 |             gate.gelu()?
[INFO] [stdout]     |             ^^^^^^^^^^^^ the `?` operator cannot be applied to type `Tensor<B>`
[INFO] [stdout]     |
[INFO] [stdout]     = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:570:48
[INFO] [stdout]     |
[INFO] [stdout] 549 | impl<B: Backend> BitNetBlock<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 570 |         let attn_output = self.dropout.forward(&attn_output)?;
[INFO] [stdout]     |                                        ------- ^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor<B>`
[INFO] [stdout]     |                                        |
[INFO] [stdout]     |                                        arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<Cpu>`
[INFO] [stdout]                found reference `&Tensor<B>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `add` found for reference `&Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:571:43
[INFO] [stdout]     |
[INFO] [stdout] 571 |         let hidden_states = hidden_states.add(&attn_output)?;
[INFO] [stdout]     |                                           ^^^ this is an associated function, not a method
[INFO] [stdout]     |
[INFO] [stdout]     = note: found the following associated functions; to be used as methods, functions must have a `self` parameter
[INFO] [stdout] note: the candidate is defined in the trait `Backend`
[INFO] [stdout]    --> crates/myllm-core/src/backend/mod.rs:58:5
[INFO] [stdout]     |
[INFO] [stdout]  58 |     fn add(a: &Self::Storage, b: &Self::Storage) -> Result<Self::Storage>;
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     = help: items from traits can only be used if the trait is in scope
[INFO] [stdout] help: use associated function syntax instead
[INFO] [stdout]     |
[INFO] [stdout] 571 -         let hidden_states = hidden_states.add(&attn_output)?;
[INFO] [stdout] 571 +         let hidden_states = Tensor<B>::add(&attn_output)?;
[INFO] [stdout]     |
[INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it
[INFO] [stdout]     |
[INFO] [stdout]  12 + use std::ops::Add;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:575:47
[INFO] [stdout]     |
[INFO] [stdout] 549 | impl<B: Backend> BitNetBlock<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 575 |         let ffn_output = self.dropout.forward(&ffn_output)?;
[INFO] [stdout]     |                                       ------- ^^^^^^^^^^^ expected `&Tensor`, found `&Tensor<B>`
[INFO] [stdout]     |                                       |
[INFO] [stdout]     |                                       arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<Cpu>`
[INFO] [stdout]                found reference `&Tensor<B>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:602:28
[INFO] [stdout]     |
[INFO] [stdout] 602 |         let embed_tokens = Embedding::new(config.vocab_size, config.hidden_size);
[INFO] [stdout]     |                            ^^^^^^^^^^^^^^ -----------------  ------------------ unexpected argument #2 of type `usize`
[INFO] [stdout]     |                                           |
[INFO] [stdout]     |                                           expected `EmbeddingConfig`, found `usize`
[INFO] [stdout]     |
[INFO] [stdout] note: associated function defined here
[INFO] [stdout]    --> crates/myllm-nn/src/embedding.rs:35:12
[INFO] [stdout]     |
[INFO] [stdout]  35 |     pub fn new(config: EmbeddingConfig) -> Self {
[INFO] [stdout]     |            ^^^
[INFO] [stdout] help: remove the extra argument
[INFO] [stdout]     |
[INFO] [stdout] 602 -         let embed_tokens = Embedding::new(config.vocab_size, config.hidden_size);
[INFO] [stdout] 602 +         let embed_tokens = Embedding::new(/* EmbeddingConfig */);
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:648:53
[INFO] [stdout]     |
[INFO] [stdout] 648 |         params.extend(prefix_params("embed_tokens", self.embed_tokens.parameters()));
[INFO] [stdout]     |                       -------------                 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:655:48
[INFO] [stdout]     |
[INFO] [stdout] 655 |         params.extend(prefix_params("lm_head", self.lm_head.parameters()));
[INFO] [stdout]     |                       -------------            ^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no function or associated item named `from_vec` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/bitnet.rs:712:13
[INFO] [stdout]     |
[INFO] [stdout] 712 |     Tensor::from_vec(mask, [seq_len, seq_len])
[INFO] [stdout]     |             ^^^^^^^^ function or associated item not found in `Tensor<_>`
[INFO] [stdout]     |
[INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions:
[INFO] [stdout]       Tensor::<B>::from_storage
[INFO] [stdout]       Tensor::<B>::from_slice
[INFO] [stdout]       Tensor::<B>::zeros
[INFO] [stdout]       Tensor::<B>::ones
[INFO] [stdout]       and 7 others
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:20:5
[INFO] [stdout]     |
[INFO] [stdout]  20 |     pub fn from_storage(storage: B::Storage) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  29 |     pub fn from_slice(data: &[f32], shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  35 |     pub fn zeros(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  40 |     pub fn ones(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] help: there is an associated function `from` with a similar name
[INFO] [stdout]    --> /rustc/59807616e1fa2540724bfbac14d7976d7e4a3860/library/core/src/convert/mod.rs:592:4
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `mul_scalar` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:289:25
[INFO] [stdout]     |
[INFO] [stdout] 289 |         scores = scores.mul_scalar(self.scale)?;
[INFO] [stdout]     |                         ^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: there is a method `add_scalar` with a similar name
[INFO] [stdout]     |
[INFO] [stdout] 289 -         scores = scores.mul_scalar(self.scale)?;
[INFO] [stdout] 289 +         scores = scores.add_scalar(self.scale)?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no function or associated item named `full` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:293:35
[INFO] [stdout]     |
[INFO] [stdout] 293 |             let neg_inf = Tensor::full(scores.shape().dims(), -1e9)?;
[INFO] [stdout]     |                                   ^^^^ function or associated item not found in `Tensor<_>`
[INFO] [stdout]     |
[INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions:
[INFO] [stdout]       Tensor::<B>::from_storage
[INFO] [stdout]       Tensor::<B>::from_slice
[INFO] [stdout]       Tensor::<B>::zeros
[INFO] [stdout]       Tensor::<B>::ones
[INFO] [stdout]       and 7 others
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:20:5
[INFO] [stdout]     |
[INFO] [stdout]  20 |     pub fn from_storage(storage: B::Storage) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  29 |     pub fn from_slice(data: &[f32], shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  35 |     pub fn zeros(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  40 |     pub fn ones(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this method takes 0 arguments but 1 argument was supplied
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:297:35
[INFO] [stdout]     |
[INFO] [stdout] 297 |         let attn_weights = scores.softmax(-1)?;
[INFO] [stdout]     |                                   ^^^^^^^ -- unexpected argument of type `{integer}`
[INFO] [stdout]     |
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:189:12
[INFO] [stdout]     |
[INFO] [stdout] 189 |     pub fn softmax(&self) -> Self {
[INFO] [stdout]     |            ^^^^^^^
[INFO] [stdout] help: remove the extra argument
[INFO] [stdout]     |
[INFO] [stdout] 297 -         let attn_weights = scores.softmax(-1)?;
[INFO] [stdout] 297 +         let attn_weights = scores.softmax()?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try`
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:297:28
[INFO] [stdout]     |
[INFO] [stdout] 297 |         let attn_weights = scores.softmax(-1)?;
[INFO] [stdout]     |                            ^^^^^^^^^^^^^^^^^^^ the `?` operator cannot be applied to type `Tensor<B>`
[INFO] [stdout]     |
[INFO] [stdout]     = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:300:47
[INFO] [stdout]     |
[INFO] [stdout] 232 | impl<B: Backend> BidirectionalAttention<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 300 |         let attn_output = attn_weights.matmul(&v)?;
[INFO] [stdout]     |                                        ------ ^^ expected `&Tensor`, found `&Tensor<B>`
[INFO] [stdout]     |                                        |
[INFO] [stdout]     |                                        arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<Cpu>`
[INFO] [stdout]                found reference `&Tensor<B>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:139:12
[INFO] [stdout]     |
[INFO] [stdout] 139 |     pub fn matmul(&self, other: &Self) -> Result<Self> {
[INFO] [stdout]     |            ^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:305:29
[INFO] [stdout]     |
[INFO] [stdout] 232 | impl<B: Backend> BidirectionalAttention<B> {
[INFO] [stdout]     |      - expected this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 305 |         self.o_proj.forward(&attn_output)
[INFO] [stdout]     |                     ------- ^^^^^^^^^^^^ expected `&Tensor<B>`, found `&Tensor`
[INFO] [stdout]     |                     |
[INFO] [stdout]     |                     arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<B>`
[INFO] [stdout]                found reference `&Tensor<Cpu>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:311:47
[INFO] [stdout]     |
[INFO] [stdout] 311 |         params.extend(prefix_params("q_proj", self.q_proj.parameters()));
[INFO] [stdout]     |                       -------------           ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:312:47
[INFO] [stdout]     |
[INFO] [stdout] 312 |         params.extend(prefix_params("k_proj", self.k_proj.parameters()));
[INFO] [stdout]     |                       -------------           ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:313:47
[INFO] [stdout]     |
[INFO] [stdout] 313 |         params.extend(prefix_params("v_proj", self.v_proj.parameters()));
[INFO] [stdout]     |                       -------------           ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:314:47
[INFO] [stdout]     |
[INFO] [stdout] 314 |         params.extend(prefix_params("o_proj", self.o_proj.parameters()));
[INFO] [stdout]     |                       -------------           ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try`
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:351:63
[INFO] [stdout]     |
[INFO] [stdout] 351 |             ActivationType::Gelu | ActivationType::GeluNew => h.gelu()?,
[INFO] [stdout]     |                                                               ^^^^^^^^^ the `?` operator cannot be applied to type `Tensor<B>`
[INFO] [stdout]     |
[INFO] [stdout]     = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try`
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:352:37
[INFO] [stdout]     |
[INFO] [stdout] 352 |             ActivationType::Relu => h.relu()?,
[INFO] [stdout]     |                                     ^^^^^^^^^ the `?` operator cannot be applied to type `Tensor<B>`
[INFO] [stdout]     |
[INFO] [stdout]     = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try`
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:353:37
[INFO] [stdout]     |
[INFO] [stdout] 353 |             ActivationType::Silu => h.silu()?,
[INFO] [stdout]     |                                     ^^^^^^^^^ the `?` operator cannot be applied to type `Tensor<B>`
[INFO] [stdout]     |
[INFO] [stdout]     = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try`
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:354:37
[INFO] [stdout]     |
[INFO] [stdout] 354 |             ActivationType::Tanh => h.tanh()?,
[INFO] [stdout]     |                                     ^^^^^^^^^ the `?` operator cannot be applied to type `Tensor<B>`
[INFO] [stdout]     |
[INFO] [stdout]     = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:358:30
[INFO] [stdout]     |
[INFO] [stdout] 327 | impl<B: Backend> EncoderFFN<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 358 |         self.dropout.forward(&h)
[INFO] [stdout]     |                      ------- ^^ expected `&Tensor`, found `&Tensor<B>`
[INFO] [stdout]     |                      |
[INFO] [stdout]     |                      arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<Cpu>`
[INFO] [stdout]                found reference `&Tensor<B>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:358:9
[INFO] [stdout]     |
[INFO] [stdout] 327 | impl<B: Backend> EncoderFFN<B> {
[INFO] [stdout]     |      - expected this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 347 |     pub fn forward(&self, hidden_states: &Tensor<B>) -> Result<Tensor<B>> {
[INFO] [stdout]     |                                                         ----------------- expected `Result<Tensor<B>, myllm_core::Error>` because of return type
[INFO] [stdout] ...
[INFO] [stdout] 358 |         self.dropout.forward(&h)
[INFO] [stdout]     |         ^^^^^^^^^^^^^^^^^^^^^^^^ expected `Result<Tensor<B>, Error>`, found `Result<Tensor, Error>`
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected enum `Result<Tensor<B>, _>`
[INFO] [stdout]                found enum `Result<Tensor<Cpu>, _>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:364:53
[INFO] [stdout]     |
[INFO] [stdout] 364 |         params.extend(prefix_params("intermediate", self.intermediate.parameters()));
[INFO] [stdout]     |                       -------------                 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:365:47
[INFO] [stdout]     |
[INFO] [stdout] 365 |         params.extend(prefix_params("output", self.output.parameters()));
[INFO] [stdout]     |                       -------------           ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:384:29
[INFO] [stdout]     |
[INFO] [stdout] 384 |             attention_norm: LayerNorm::new(config.hidden_size, config.layer_norm_eps),
[INFO] [stdout]     |                             ^^^^^^^^^^^^^^ ------------------  --------------------- unexpected argument #2 of type `f32`
[INFO] [stdout]     |                                            |
[INFO] [stdout]     |                                            expected `LayerNormConfig`, found `usize`
[INFO] [stdout]     |
[INFO] [stdout] note: associated function defined here
[INFO] [stdout]    --> crates/myllm-nn/src/norm.rs:50:12
[INFO] [stdout]     |
[INFO] [stdout]  50 |     pub fn new(config: LayerNormConfig) -> Self {
[INFO] [stdout]     |            ^^^
[INFO] [stdout] help: remove the extra argument
[INFO] [stdout]     |
[INFO] [stdout] 384 -             attention_norm: LayerNorm::new(config.hidden_size, config.layer_norm_eps),
[INFO] [stdout] 384 +             attention_norm: LayerNorm::new(/* LayerNormConfig */),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:386:23
[INFO] [stdout]     |
[INFO] [stdout] 386 |             ffn_norm: LayerNorm::new(config.hidden_size, config.layer_norm_eps),
[INFO] [stdout]     |                       ^^^^^^^^^^^^^^ ------------------  --------------------- unexpected argument #2 of type `f32`
[INFO] [stdout]     |                                      |
[INFO] [stdout]     |                                      expected `LayerNormConfig`, found `usize`
[INFO] [stdout]     |
[INFO] [stdout] note: associated function defined here
[INFO] [stdout]    --> crates/myllm-nn/src/norm.rs:50:12
[INFO] [stdout]     |
[INFO] [stdout]  50 |     pub fn new(config: LayerNormConfig) -> Self {
[INFO] [stdout]     |            ^^^
[INFO] [stdout] help: remove the extra argument
[INFO] [stdout]     |
[INFO] [stdout] 386 -             ffn_norm: LayerNorm::new(config.hidden_size, config.layer_norm_eps),
[INFO] [stdout] 386 +             ffn_norm: LayerNorm::new(/* LayerNormConfig */),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:398:48
[INFO] [stdout]     |
[INFO] [stdout] 379 | impl<B: Backend> EncoderLayer<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 398 |         let attn_output = self.dropout.forward(&attn_output)?;
[INFO] [stdout]     |                                        ------- ^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor<B>`
[INFO] [stdout]     |                                        |
[INFO] [stdout]     |                                        arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<Cpu>`
[INFO] [stdout]                found reference `&Tensor<B>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `add` found for reference `&Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:399:43
[INFO] [stdout]     |
[INFO] [stdout] 399 |         let hidden_states = hidden_states.add(&attn_output)?;
[INFO] [stdout]     |                                           ^^^ this is an associated function, not a method
[INFO] [stdout]     |
[INFO] [stdout]     = note: found the following associated functions; to be used as methods, functions must have a `self` parameter
[INFO] [stdout] note: the candidate is defined in the trait `Backend`
[INFO] [stdout]    --> crates/myllm-core/src/backend/mod.rs:58:5
[INFO] [stdout]     |
[INFO] [stdout]  58 |     fn add(a: &Self::Storage, b: &Self::Storage) -> Result<Self::Storage>;
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     = help: items from traits can only be used if the trait is in scope
[INFO] [stdout] help: use associated function syntax instead
[INFO] [stdout]     |
[INFO] [stdout] 399 -         let hidden_states = hidden_states.add(&attn_output)?;
[INFO] [stdout] 399 +         let hidden_states = Tensor<B>::add(&attn_output)?;
[INFO] [stdout]     |
[INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it
[INFO] [stdout]     |
[INFO] [stdout]   9 + use std::ops::Add;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `add` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:403:43
[INFO] [stdout]     |
[INFO] [stdout] 403 |         let hidden_states = hidden_states.add(&ffn_output)?;
[INFO] [stdout]     |                                           ^^^ this is an associated function, not a method
[INFO] [stdout]     |
[INFO] [stdout]    --> /rustc/59807616e1fa2540724bfbac14d7976d7e4a3860/library/core/src/ops/arith.rs:92:7
[INFO] [stdout]     |
[INFO] [stdout]     = note: the method is available for `Tensor<B>` here
[INFO] [stdout]     |
[INFO] [stdout]     = note: found the following associated functions; to be used as methods, functions must have a `self` parameter
[INFO] [stdout] note: the candidate is defined in the trait `Backend`
[INFO] [stdout]    --> crates/myllm-core/src/backend/mod.rs:58:5
[INFO] [stdout]     |
[INFO] [stdout]  58 |     fn add(a: &Self::Storage, b: &Self::Storage) -> Result<Self::Storage>;
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     = help: items from traits can only be used if the trait is in scope
[INFO] [stdout] help: use associated function syntax instead
[INFO] [stdout]     |
[INFO] [stdout] 403 -         let hidden_states = hidden_states.add(&ffn_output)?;
[INFO] [stdout] 403 +         let hidden_states = Tensor<B>::add(&ffn_output)?;
[INFO] [stdout]     |
[INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it
[INFO] [stdout]     |
[INFO] [stdout]   9 + use std::ops::Add;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:411:55
[INFO] [stdout]     |
[INFO] [stdout] 411 |         params.extend(prefix_params("attention_norm", self.attention_norm.parameters()));
[INFO] [stdout]     |                       -------------                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:413:49
[INFO] [stdout]     |
[INFO] [stdout] 413 |         params.extend(prefix_params("ffn_norm", self.ffn_norm.parameters()));
[INFO] [stdout]     |                       -------------             ^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:431:30
[INFO] [stdout]     |
[INFO] [stdout] 431 |             word_embeddings: Embedding::new(config.vocab_size, config.hidden_size),
[INFO] [stdout]     |                              ^^^^^^^^^^^^^^ -----------------  ------------------ unexpected argument #2 of type `usize`
[INFO] [stdout]     |                                             |
[INFO] [stdout]     |                                             expected `EmbeddingConfig`, found `usize`
[INFO] [stdout]     |
[INFO] [stdout] note: associated function defined here
[INFO] [stdout]    --> crates/myllm-nn/src/embedding.rs:35:12
[INFO] [stdout]     |
[INFO] [stdout]  35 |     pub fn new(config: EmbeddingConfig) -> Self {
[INFO] [stdout]     |            ^^^
[INFO] [stdout] help: remove the extra argument
[INFO] [stdout]     |
[INFO] [stdout] 431 -             word_embeddings: Embedding::new(config.vocab_size, config.hidden_size),
[INFO] [stdout] 431 +             word_embeddings: Embedding::new(/* EmbeddingConfig */),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:432:34
[INFO] [stdout]     |
[INFO] [stdout] 432 |             position_embeddings: Embedding::new(config.max_position_embeddings, config.hidden_size),
[INFO] [stdout]     |                                  ^^^^^^^^^^^^^^ ------------------------------  ------------------ unexpected argument #2 of type `usize`
[INFO] [stdout]     |                                                 |
[INFO] [stdout]     |                                                 expected `EmbeddingConfig`, found `usize`
[INFO] [stdout]     |
[INFO] [stdout] note: associated function defined here
[INFO] [stdout]    --> crates/myllm-nn/src/embedding.rs:35:12
[INFO] [stdout]     |
[INFO] [stdout]  35 |     pub fn new(config: EmbeddingConfig) -> Self {
[INFO] [stdout]     |            ^^^
[INFO] [stdout] help: remove the extra argument
[INFO] [stdout]     |
[INFO] [stdout] 432 -             position_embeddings: Embedding::new(config.max_position_embeddings, config.hidden_size),
[INFO] [stdout] 432 +             position_embeddings: Embedding::new(/* EmbeddingConfig */),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:433:36
[INFO] [stdout]     |
[INFO] [stdout] 433 |             token_type_embeddings: Embedding::new(config.type_vocab_size, config.hidden_size),
[INFO] [stdout]     |                                    ^^^^^^^^^^^^^^ ----------------------  ------------------ unexpected argument #2 of type `usize`
[INFO] [stdout]     |                                                   |
[INFO] [stdout]     |                                                   expected `EmbeddingConfig`, found `usize`
[INFO] [stdout]     |
[INFO] [stdout] note: associated function defined here
[INFO] [stdout]    --> crates/myllm-nn/src/embedding.rs:35:12
[INFO] [stdout]     |
[INFO] [stdout]  35 |     pub fn new(config: EmbeddingConfig) -> Self {
[INFO] [stdout]     |            ^^^
[INFO] [stdout] help: remove the extra argument
[INFO] [stdout]     |
[INFO] [stdout] 433 -             token_type_embeddings: Embedding::new(config.type_vocab_size, config.hidden_size),
[INFO] [stdout] 433 +             token_type_embeddings: Embedding::new(/* EmbeddingConfig */),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:434:25
[INFO] [stdout]     |
[INFO] [stdout] 434 |             layer_norm: LayerNorm::new(config.hidden_size, config.layer_norm_eps),
[INFO] [stdout]     |                         ^^^^^^^^^^^^^^ ------------------  --------------------- unexpected argument #2 of type `f32`
[INFO] [stdout]     |                                        |
[INFO] [stdout]     |                                        expected `LayerNormConfig`, found `usize`
[INFO] [stdout]     |
[INFO] [stdout] note: associated function defined here
[INFO] [stdout]    --> crates/myllm-nn/src/norm.rs:50:12
[INFO] [stdout]     |
[INFO] [stdout]  50 |     pub fn new(config: LayerNormConfig) -> Self {
[INFO] [stdout]     |            ^^^
[INFO] [stdout] help: remove the extra argument
[INFO] [stdout]     |
[INFO] [stdout] 434 -             layer_norm: LayerNorm::new(config.hidden_size, config.layer_norm_eps),
[INFO] [stdout] 434 +             layer_norm: LayerNorm::new(/* LayerNormConfig */),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no function or associated item named `from_vec` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:454:21
[INFO] [stdout]     |
[INFO] [stdout] 454 |             Tensor::from_vec(positions, [1, seq_len])?
[INFO] [stdout]     |                     ^^^^^^^^ function or associated item not found in `Tensor<_>`
[INFO] [stdout]     |
[INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions:
[INFO] [stdout]       Tensor::<B>::from_storage
[INFO] [stdout]       Tensor::<B>::from_slice
[INFO] [stdout]       Tensor::<B>::zeros
[INFO] [stdout]       Tensor::<B>::ones
[INFO] [stdout]       and 7 others
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:20:5
[INFO] [stdout]     |
[INFO] [stdout]  20 |     pub fn from_storage(storage: B::Storage) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  29 |     pub fn from_slice(data: &[f32], shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  35 |     pub fn zeros(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  40 |     pub fn ones(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] help: there is an associated function `from` with a similar name
[INFO] [stdout]    --> /rustc/59807616e1fa2540724bfbac14d7976d7e4a3860/library/core/src/convert/mod.rs:592:4
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try`
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:461:13
[INFO] [stdout]     |
[INFO] [stdout] 461 |             Tensor::zeros([input_ids.shape().dim(0), seq_len, word_embeds.shape().dim(2)])?
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ the `?` operator cannot be applied to type `Tensor<_>`
[INFO] [stdout]     |
[INFO] [stdout]     = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor<_>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `add` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:464:38
[INFO] [stdout]     |
[INFO] [stdout] 464 |         let embeddings = word_embeds.add(&position_embeds)?.add(&token_type_embeds)?;
[INFO] [stdout]     |                                      ^^^ this is an associated function, not a method
[INFO] [stdout]     |
[INFO] [stdout]    --> /rustc/59807616e1fa2540724bfbac14d7976d7e4a3860/library/core/src/ops/arith.rs:92:7
[INFO] [stdout]     |
[INFO] [stdout]     = note: the method is available for `Tensor<B>` here
[INFO] [stdout]     |
[INFO] [stdout]     = note: found the following associated functions; to be used as methods, functions must have a `self` parameter
[INFO] [stdout] note: the candidate is defined in the trait `Backend`
[INFO] [stdout]    --> crates/myllm-core/src/backend/mod.rs:58:5
[INFO] [stdout]     |
[INFO] [stdout]  58 |     fn add(a: &Self::Storage, b: &Self::Storage) -> Result<Self::Storage>;
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     = help: items from traits can only be used if the trait is in scope
[INFO] [stdout] help: use associated function syntax instead
[INFO] [stdout]     |
[INFO] [stdout] 464 -         let embeddings = word_embeds.add(&position_embeds)?.add(&token_type_embeds)?;
[INFO] [stdout] 464 +         let embeddings = Tensor<B>::add(&position_embeds)?.add(&token_type_embeds)?;
[INFO] [stdout]     |
[INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it
[INFO] [stdout]     |
[INFO] [stdout]   9 + use std::ops::Add;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:466:30
[INFO] [stdout]     |
[INFO] [stdout] 427 | impl<B: Backend> BertEmbeddings<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 466 |         self.dropout.forward(&embeddings)
[INFO] [stdout]     |                      ------- ^^^^^^^^^^^ expected `&Tensor`, found `&Tensor<B>`
[INFO] [stdout]     |                      |
[INFO] [stdout]     |                      arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<Cpu>`
[INFO] [stdout]                found reference `&Tensor<B>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:466:9
[INFO] [stdout]     |
[INFO] [stdout] 427 | impl<B: Backend> BertEmbeddings<B> {
[INFO] [stdout]     |      - expected this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 445 |     ) -> Result<Tensor<B>> {
[INFO] [stdout]     |          ----------------- expected `Result<Tensor<B>, myllm_core::Error>` because of return type
[INFO] [stdout] ...
[INFO] [stdout] 466 |         self.dropout.forward(&embeddings)
[INFO] [stdout]     |         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `Result<Tensor<B>, Error>`, found `Result<Tensor, Error>`
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected enum `Result<Tensor<B>, _>`
[INFO] [stdout]                found enum `Result<Tensor<Cpu>, _>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:472:56
[INFO] [stdout]     |
[INFO] [stdout] 472 |         params.extend(prefix_params("word_embeddings", self.word_embeddings.parameters()));
[INFO] [stdout]     |                       -------------                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:473:60
[INFO] [stdout]     |
[INFO] [stdout] 473 |         params.extend(prefix_params("position_embeddings", self.position_embeddings.parameters()));
[INFO] [stdout]     |                       -------------                        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:474:62
[INFO] [stdout]     |
[INFO] [stdout] 474 |         params.extend(prefix_params("token_type_embeddings", self.token_type_embeddings.parameters()));
[INFO] [stdout]     |                       -------------                          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:475:51
[INFO] [stdout]     |
[INFO] [stdout] 475 |         params.extend(prefix_params("layer_norm", self.layer_norm.parameters()));
[INFO] [stdout]     |                       -------------               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `squeeze` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:551:21
[INFO] [stdout]     |
[INFO] [stdout] 551 |                 cls.squeeze(1)
[INFO] [stdout]     |                     ^^^^^^^ method not found in `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `squeeze` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:571:70
[INFO] [stdout]     |
[INFO] [stdout] 571 |                 let first = output.last_hidden_state.slice(1, 0, 1)?.squeeze(1)?;
[INFO] [stdout]     |                                                                      ^^^^^^^ method not found in `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `squeeze` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:573:87
[INFO] [stdout]     |
[INFO] [stdout] 573 |                 let last = output.last_hidden_state.slice(1, last_idx, last_idx + 1)?.squeeze(1)?;
[INFO] [stdout]     |                                                                                       ^^^^^^^ method not found in `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0600]: cannot apply unary operator `-` to type `usize`
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:574:47
[INFO] [stdout]     |
[INFO] [stdout] 574 |                 Tensor::cat(&[&first, &last], -1)
[INFO] [stdout]     |                                               ^^ cannot apply unary operator `-`
[INFO] [stdout]     |
[INFO] [stdout]     = note: unsigned values cannot be negated
[INFO] [stdout] help: you may have meant the maximum value of `usize`
[INFO] [stdout]     |
[INFO] [stdout] 574 -                 Tensor::cat(&[&first, &last], -1)
[INFO] [stdout] 574 +                 Tensor::cat(&[&first, &last], usize::MAX)
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `squeeze` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:631:35
[INFO] [stdout]     |
[INFO] [stdout] 631 |         let cls_token = cls_token.squeeze(1)?;
[INFO] [stdout]     |                                   ^^^^^^^ method not found in `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:633:9
[INFO] [stdout]     |
[INFO] [stdout] 629 |     pub fn forward(&self, hidden_states: &Tensor<B>) -> Result<Tensor<B>> {
[INFO] [stdout]     |                                                         ----------------- expected `Result<Tensor<B>, myllm_core::Error>` because of return type
[INFO] [stdout] ...
[INFO] [stdout] 633 |         pooled.tanh()
[INFO] [stdout]     |         ^^^^^^^^^^^^^ expected `Result<Tensor<B>, Error>`, found `Tensor<B>`
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected enum `Result<Tensor<_>, myllm_core::Error>`
[INFO] [stdout]              found struct `Tensor<_>`
[INFO] [stdout] help: try wrapping the expression in `Ok`
[INFO] [stdout]     |
[INFO] [stdout] 633 |         Ok(pooled.tanh())
[INFO] [stdout]     |         +++             +
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:638:32
[INFO] [stdout]     |
[INFO] [stdout] 638 |         prefix_params("dense", self.dense.parameters())
[INFO] [stdout]     |         -------------          ^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<B>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |         |
[INFO] [stdout]     |         arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<B>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:661:44
[INFO] [stdout]     |
[INFO] [stdout] 648 | impl<B: Backend> BertClassificationHead<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 661 |         let dropped = self.dropout.forward(pooled_output)?;
[INFO] [stdout]     |                                    ------- ^^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor<B>`
[INFO] [stdout]     |                                    |
[INFO] [stdout]     |                                    arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<Cpu>`
[INFO] [stdout]                found reference `&Tensor<B>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:662:33
[INFO] [stdout]     |
[INFO] [stdout] 648 | impl<B: Backend> BertClassificationHead<B> {
[INFO] [stdout]     |      - expected this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 662 |         self.classifier.forward(&dropped)
[INFO] [stdout]     |                         ------- ^^^^^^^^ expected `&Tensor<B>`, found `&Tensor`
[INFO] [stdout]     |                         |
[INFO] [stdout]     |                         arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<B>`
[INFO] [stdout]                found reference `&Tensor<Cpu>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:667:37
[INFO] [stdout]     |
[INFO] [stdout] 667 |         prefix_params("classifier", self.classifier.parameters())
[INFO] [stdout]     |         -------------               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<B>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |         |
[INFO] [stdout]     |         arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<B>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:690:44
[INFO] [stdout]     |
[INFO] [stdout] 677 | impl<B: Backend> BertTokenClassificationHead<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 690 |         let dropped = self.dropout.forward(sequence_output)?;
[INFO] [stdout]     |                                    ------- ^^^^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor<B>`
[INFO] [stdout]     |                                    |
[INFO] [stdout]     |                                    arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<Cpu>`
[INFO] [stdout]                found reference `&Tensor<B>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:691:33
[INFO] [stdout]     |
[INFO] [stdout] 677 | impl<B: Backend> BertTokenClassificationHead<B> {
[INFO] [stdout]     |      - expected this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 691 |         self.classifier.forward(&dropped)
[INFO] [stdout]     |                         ------- ^^^^^^^^ expected `&Tensor<B>`, found `&Tensor`
[INFO] [stdout]     |                         |
[INFO] [stdout]     |                         arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<B>`
[INFO] [stdout]                found reference `&Tensor<Cpu>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:696:37
[INFO] [stdout]     |
[INFO] [stdout] 696 |         prefix_params("classifier", self.classifier.parameters())
[INFO] [stdout]     |         -------------               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<B>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |         |
[INFO] [stdout]     |         arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<B>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `unsqueeze` found for reference `&Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:796:40
[INFO] [stdout]     |
[INFO] [stdout] 796 |     let mask_expanded = attention_mask.unsqueeze(-1)?;
[INFO] [stdout]     |                                        ^^^^^^^^^ method not found in `&Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `mul` found for reference `&Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:798:32
[INFO] [stdout]     |
[INFO] [stdout] 798 |     let masked = hidden_states.mul(&mask_expanded)?;
[INFO] [stdout]     |                                ^^^ this is an associated function, not a method
[INFO] [stdout]     |
[INFO] [stdout]     = note: found the following associated functions; to be used as methods, functions must have a `self` parameter
[INFO] [stdout] note: the candidate is defined in the trait `Backend`
[INFO] [stdout]    --> crates/myllm-core/src/backend/mod.rs:64:5
[INFO] [stdout]     |
[INFO] [stdout]  64 |     fn mul(a: &Self::Storage, b: &Self::Storage) -> Result<Self::Storage>;
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     = help: items from traits can only be used if the trait is in scope
[INFO] [stdout] help: use associated function syntax instead
[INFO] [stdout]     |
[INFO] [stdout] 798 -     let masked = hidden_states.mul(&mask_expanded)?;
[INFO] [stdout] 798 +     let masked = Tensor<B>::mul(&mask_expanded)?;
[INFO] [stdout]     |
[INFO] [stdout] help: trait `Mul` which provides `mul` is implemented but not in scope; perhaps you want to import it
[INFO] [stdout]     |
[INFO] [stdout]   9 + use std::ops::Mul;
[INFO] [stdout]     |
[INFO] [stdout] help: there is a method `matmul` with a similar name
[INFO] [stdout]     |
[INFO] [stdout] 798 |     let masked = hidden_states.matmul(&mask_expanded)?;
[INFO] [stdout]     |                                +++
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:188:38
[INFO] [stdout]     |
[INFO] [stdout] 188 |             relative_attention_bias: Embedding::new(
[INFO] [stdout]     |                                      ^^^^^^^^^^^^^^
[INFO] [stdout] 189 |                 config.relative_position_buckets,
[INFO] [stdout]     |                 -------------------------------- expected `EmbeddingConfig`, found `usize`
[INFO] [stdout] 190 |                 config.num_heads,
[INFO] [stdout]     |                 ---------------- unexpected argument #2 of type `usize`
[INFO] [stdout]     |
[INFO] [stdout] note: associated function defined here
[INFO] [stdout]    --> crates/myllm-nn/src/embedding.rs:35:12
[INFO] [stdout]     |
[INFO] [stdout]  35 |     pub fn new(config: EmbeddingConfig) -> Self {
[INFO] [stdout]     |            ^^^
[INFO] [stdout] help: remove the extra argument
[INFO] [stdout]     |
[INFO] [stdout] 189 -                 config.relative_position_buckets,
[INFO] [stdout] 190 -                 config.num_heads,
[INFO] [stdout] 189 +                 /* EmbeddingConfig */,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no function or associated item named `from_vec` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:256:36
[INFO] [stdout]     |
[INFO] [stdout] 256 |         let position_ids = Tensor::from_vec(position_ids, [query_length, key_length])?;
[INFO] [stdout]     |                                    ^^^^^^^^ function or associated item not found in `Tensor<_>`
[INFO] [stdout]     |
[INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions:
[INFO] [stdout]       Tensor::<B>::from_storage
[INFO] [stdout]       Tensor::<B>::from_slice
[INFO] [stdout]       Tensor::<B>::zeros
[INFO] [stdout]       Tensor::<B>::ones
[INFO] [stdout]       and 7 others
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:20:5
[INFO] [stdout]     |
[INFO] [stdout]  20 |     pub fn from_storage(storage: B::Storage) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  29 |     pub fn from_slice(data: &[f32], shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  35 |     pub fn zeros(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  40 |     pub fn ones(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] help: there is an associated function `from` with a similar name
[INFO] [stdout]    --> /rustc/59807616e1fa2540724bfbac14d7976d7e4a3860/library/core/src/convert/mod.rs:592:4
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `unsqueeze` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:259:35
[INFO] [stdout]     |
[INFO] [stdout] 259 |         bias.permute(&[2, 0, 1])?.unsqueeze(0)
[INFO] [stdout]     |                                   ^^^^^^^^^ method not found in `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:264:9
[INFO] [stdout]     |
[INFO] [stdout] 263 |     pub fn parameters(&self) -> HashMap<String, Tensor<B>> {
[INFO] [stdout]     |                                 -------------------------- expected `HashMap<String, Tensor<B>>` because of return type
[INFO] [stdout] 264 |         self.relative_attention_bias.parameters()
[INFO] [stdout]     |         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<B>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<B>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `mul_scalar` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:330:25
[INFO] [stdout]     |
[INFO] [stdout] 330 |         scores = scores.mul_scalar(self.scale)?;
[INFO] [stdout]     |                         ^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: there is a method `add_scalar` with a similar name
[INFO] [stdout]     |
[INFO] [stdout] 330 -         scores = scores.mul_scalar(self.scale)?;
[INFO] [stdout] 330 +         scores = scores.add_scalar(self.scale)?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `add` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:333:29
[INFO] [stdout]     |
[INFO] [stdout] 333 |             scores = scores.add(bias)?;
[INFO] [stdout]     |                             ^^^ this is an associated function, not a method
[INFO] [stdout]     |
[INFO] [stdout]    --> /rustc/59807616e1fa2540724bfbac14d7976d7e4a3860/library/core/src/ops/arith.rs:92:7
[INFO] [stdout]     |
[INFO] [stdout]     = note: the method is available for `Tensor<B>` here
[INFO] [stdout]     |
[INFO] [stdout]     = note: found the following associated functions; to be used as methods, functions must have a `self` parameter
[INFO] [stdout] note: the candidate is defined in the trait `Backend`
[INFO] [stdout]    --> crates/myllm-core/src/backend/mod.rs:58:5
[INFO] [stdout]     |
[INFO] [stdout]  58 |     fn add(a: &Self::Storage, b: &Self::Storage) -> Result<Self::Storage>;
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     = help: items from traits can only be used if the trait is in scope
[INFO] [stdout] help: use associated function syntax instead
[INFO] [stdout]     |
[INFO] [stdout] 333 -             scores = scores.add(bias)?;
[INFO] [stdout] 333 +             scores = Tensor<B>::add(bias)?;
[INFO] [stdout]     |
[INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it
[INFO] [stdout]     |
[INFO] [stdout]   9 + use std::ops::Add;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no function or associated item named `full` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:338:35
[INFO] [stdout]     |
[INFO] [stdout] 338 |             let neg_inf = Tensor::full(scores.shape().dims(), -1e9)?;
[INFO] [stdout]     |                                   ^^^^ function or associated item not found in `Tensor<_>`
[INFO] [stdout]     |
[INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions:
[INFO] [stdout]       Tensor::<B>::from_storage
[INFO] [stdout]       Tensor::<B>::from_slice
[INFO] [stdout]       Tensor::<B>::zeros
[INFO] [stdout]       Tensor::<B>::ones
[INFO] [stdout]       and 7 others
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:20:5
[INFO] [stdout]     |
[INFO] [stdout]  20 |     pub fn from_storage(storage: B::Storage) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  29 |     pub fn from_slice(data: &[f32], shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  35 |     pub fn zeros(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  40 |     pub fn ones(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this method takes 0 arguments but 1 argument was supplied
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:342:35
[INFO] [stdout]     |
[INFO] [stdout] 342 |         let attn_weights = scores.softmax(-1)?;
[INFO] [stdout]     |                                   ^^^^^^^ -- unexpected argument of type `{integer}`
[INFO] [stdout]     |
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:189:12
[INFO] [stdout]     |
[INFO] [stdout] 189 |     pub fn softmax(&self) -> Self {
[INFO] [stdout]     |            ^^^^^^^
[INFO] [stdout] help: remove the extra argument
[INFO] [stdout]     |
[INFO] [stdout] 342 -         let attn_weights = scores.softmax(-1)?;
[INFO] [stdout] 342 +         let attn_weights = scores.softmax()?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try`
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:342:28
[INFO] [stdout]     |
[INFO] [stdout] 342 |         let attn_weights = scores.softmax(-1)?;
[INFO] [stdout]     |                            ^^^^^^^^^^^^^^^^^^^ the `?` operator cannot be applied to type `Tensor<B>`
[INFO] [stdout]     |
[INFO] [stdout]     = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:345:47
[INFO] [stdout]     |
[INFO] [stdout] 280 | impl<B: Backend> EncoderSelfAttention<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 345 |         let attn_output = attn_weights.matmul(&v)?;
[INFO] [stdout]     |                                        ------ ^^ expected `&Tensor`, found `&Tensor<B>`
[INFO] [stdout]     |                                        |
[INFO] [stdout]     |                                        arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<Cpu>`
[INFO] [stdout]                found reference `&Tensor<B>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:139:12
[INFO] [stdout]     |
[INFO] [stdout] 139 |     pub fn matmul(&self, other: &Self) -> Result<Self> {
[INFO] [stdout]     |            ^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:350:29
[INFO] [stdout]     |
[INFO] [stdout] 280 | impl<B: Backend> EncoderSelfAttention<B> {
[INFO] [stdout]     |      - expected this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 350 |         self.o_proj.forward(&attn_output)
[INFO] [stdout]     |                     ------- ^^^^^^^^^^^^ expected `&Tensor<B>`, found `&Tensor`
[INFO] [stdout]     |                     |
[INFO] [stdout]     |                     arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<B>`
[INFO] [stdout]                found reference `&Tensor<Cpu>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:356:47
[INFO] [stdout]     |
[INFO] [stdout] 356 |         params.extend(prefix_params("q_proj", self.q_proj.parameters()));
[INFO] [stdout]     |                       -------------           ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:357:47
[INFO] [stdout]     |
[INFO] [stdout] 357 |         params.extend(prefix_params("k_proj", self.k_proj.parameters()));
[INFO] [stdout]     |                       -------------           ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:358:47
[INFO] [stdout]     |
[INFO] [stdout] 358 |         params.extend(prefix_params("v_proj", self.v_proj.parameters()));
[INFO] [stdout]     |                       -------------           ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:359:47
[INFO] [stdout]     |
[INFO] [stdout] 359 |         params.extend(prefix_params("o_proj", self.o_proj.parameters()));
[INFO] [stdout]     |                       -------------           ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `mul_scalar` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:438:25
[INFO] [stdout]     |
[INFO] [stdout] 438 |         scores = scores.mul_scalar(self.scale)?;
[INFO] [stdout]     |                         ^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: there is a method `add_scalar` with a similar name
[INFO] [stdout]     |
[INFO] [stdout] 438 -         scores = scores.mul_scalar(self.scale)?;
[INFO] [stdout] 438 +         scores = scores.add_scalar(self.scale)?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `add` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:441:29
[INFO] [stdout]     |
[INFO] [stdout] 441 |             scores = scores.add(bias)?;
[INFO] [stdout]     |                             ^^^ this is an associated function, not a method
[INFO] [stdout]     |
[INFO] [stdout]    --> /rustc/59807616e1fa2540724bfbac14d7976d7e4a3860/library/core/src/ops/arith.rs:92:7
[INFO] [stdout]     |
[INFO] [stdout]     = note: the method is available for `Tensor<B>` here
[INFO] [stdout]     |
[INFO] [stdout]     = note: found the following associated functions; to be used as methods, functions must have a `self` parameter
[INFO] [stdout] note: the candidate is defined in the trait `Backend`
[INFO] [stdout]    --> crates/myllm-core/src/backend/mod.rs:58:5
[INFO] [stdout]     |
[INFO] [stdout]  58 |     fn add(a: &Self::Storage, b: &Self::Storage) -> Result<Self::Storage>;
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     = help: items from traits can only be used if the trait is in scope
[INFO] [stdout] help: use associated function syntax instead
[INFO] [stdout]     |
[INFO] [stdout] 441 -             scores = scores.add(bias)?;
[INFO] [stdout] 441 +             scores = Tensor<B>::add(bias)?;
[INFO] [stdout]     |
[INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it
[INFO] [stdout]     |
[INFO] [stdout]   9 + use std::ops::Add;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `unsqueeze` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:445:39
[INFO] [stdout]     |
[INFO] [stdout] 445 |         let causal_mask = causal_mask.unsqueeze(0)?.unsqueeze(0)?;
[INFO] [stdout]     |                                       ^^^^^^^^^ method not found in `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no function or associated item named `full` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:446:31
[INFO] [stdout]     |
[INFO] [stdout] 446 |         let neg_inf = Tensor::full(scores.shape().dims(), -1e9)?;
[INFO] [stdout]     |                               ^^^^ function or associated item not found in `Tensor<_>`
[INFO] [stdout]     |
[INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions:
[INFO] [stdout]       Tensor::<B>::from_storage
[INFO] [stdout]       Tensor::<B>::from_slice
[INFO] [stdout]       Tensor::<B>::zeros
[INFO] [stdout]       Tensor::<B>::ones
[INFO] [stdout]       and 7 others
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:20:5
[INFO] [stdout]     |
[INFO] [stdout]  20 |     pub fn from_storage(storage: B::Storage) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  29 |     pub fn from_slice(data: &[f32], shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  35 |     pub fn zeros(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  40 |     pub fn ones(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this method takes 0 arguments but 1 argument was supplied
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:449:35
[INFO] [stdout]     |
[INFO] [stdout] 449 |         let attn_weights = scores.softmax(-1)?;
[INFO] [stdout]     |                                   ^^^^^^^ -- unexpected argument of type `{integer}`
[INFO] [stdout]     |
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:189:12
[INFO] [stdout]     |
[INFO] [stdout] 189 |     pub fn softmax(&self) -> Self {
[INFO] [stdout]     |            ^^^^^^^
[INFO] [stdout] help: remove the extra argument
[INFO] [stdout]     |
[INFO] [stdout] 449 -         let attn_weights = scores.softmax(-1)?;
[INFO] [stdout] 449 +         let attn_weights = scores.softmax()?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try`
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:449:28
[INFO] [stdout]     |
[INFO] [stdout] 449 |         let attn_weights = scores.softmax(-1)?;
[INFO] [stdout]     |                            ^^^^^^^^^^^^^^^^^^^ the `?` operator cannot be applied to type `Tensor<B>`
[INFO] [stdout]     |
[INFO] [stdout]     = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:452:47
[INFO] [stdout]     |
[INFO] [stdout] 376 | impl<B: Backend> DecoderSelfAttention<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 452 |         let attn_output = attn_weights.matmul(&v)?;
[INFO] [stdout]     |                                        ------ ^^ expected `&Tensor`, found `&Tensor<B>`
[INFO] [stdout]     |                                        |
[INFO] [stdout]     |                                        arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<Cpu>`
[INFO] [stdout]                found reference `&Tensor<B>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:139:12
[INFO] [stdout]     |
[INFO] [stdout] 139 |     pub fn matmul(&self, other: &Self) -> Result<Self> {
[INFO] [stdout]     |            ^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:457:42
[INFO] [stdout]     |
[INFO] [stdout] 376 | impl<B: Backend> DecoderSelfAttention<B> {
[INFO] [stdout]     |      - expected this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 457 |         let output = self.o_proj.forward(&attn_output)?;
[INFO] [stdout]     |                                  ------- ^^^^^^^^^^^^ expected `&Tensor<B>`, found `&Tensor`
[INFO] [stdout]     |                                  |
[INFO] [stdout]     |                                  arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<B>`
[INFO] [stdout]                found reference `&Tensor<Cpu>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:464:47
[INFO] [stdout]     |
[INFO] [stdout] 464 |         params.extend(prefix_params("q_proj", self.q_proj.parameters()));
[INFO] [stdout]     |                       -------------           ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:465:47
[INFO] [stdout]     |
[INFO] [stdout] 465 |         params.extend(prefix_params("k_proj", self.k_proj.parameters()));
[INFO] [stdout]     |                       -------------           ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:466:47
[INFO] [stdout]     |
[INFO] [stdout] 466 |         params.extend(prefix_params("v_proj", self.v_proj.parameters()));
[INFO] [stdout]     |                       -------------           ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:467:47
[INFO] [stdout]     |
[INFO] [stdout] 467 |         params.extend(prefix_params("o_proj", self.o_proj.parameters()));
[INFO] [stdout]     |                       -------------           ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `mul_scalar` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:551:25
[INFO] [stdout]     |
[INFO] [stdout] 551 |         scores = scores.mul_scalar(self.scale)?;
[INFO] [stdout]     |                         ^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: there is a method `add_scalar` with a similar name
[INFO] [stdout]     |
[INFO] [stdout] 551 -         scores = scores.mul_scalar(self.scale)?;
[INFO] [stdout] 551 +         scores = scores.add_scalar(self.scale)?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no function or associated item named `full` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:555:35
[INFO] [stdout]     |
[INFO] [stdout] 555 |             let neg_inf = Tensor::full(scores.shape().dims(), -1e9)?;
[INFO] [stdout]     |                                   ^^^^ function or associated item not found in `Tensor<_>`
[INFO] [stdout]     |
[INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions:
[INFO] [stdout]       Tensor::<B>::from_storage
[INFO] [stdout]       Tensor::<B>::from_slice
[INFO] [stdout]       Tensor::<B>::zeros
[INFO] [stdout]       Tensor::<B>::ones
[INFO] [stdout]       and 7 others
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:20:5
[INFO] [stdout]     |
[INFO] [stdout]  20 |     pub fn from_storage(storage: B::Storage) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  29 |     pub fn from_slice(data: &[f32], shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  35 |     pub fn zeros(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  40 |     pub fn ones(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this method takes 0 arguments but 1 argument was supplied
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:559:35
[INFO] [stdout]     |
[INFO] [stdout] 559 |         let attn_weights = scores.softmax(-1)?;
[INFO] [stdout]     |                                   ^^^^^^^ -- unexpected argument of type `{integer}`
[INFO] [stdout]     |
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:189:12
[INFO] [stdout]     |
[INFO] [stdout] 189 |     pub fn softmax(&self) -> Self {
[INFO] [stdout]     |            ^^^^^^^
[INFO] [stdout] help: remove the extra argument
[INFO] [stdout]     |
[INFO] [stdout] 559 -         let attn_weights = scores.softmax(-1)?;
[INFO] [stdout] 559 +         let attn_weights = scores.softmax()?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try`
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:559:28
[INFO] [stdout]     |
[INFO] [stdout] 559 |         let attn_weights = scores.softmax(-1)?;
[INFO] [stdout]     |                            ^^^^^^^^^^^^^^^^^^^ the `?` operator cannot be applied to type `Tensor<B>`
[INFO] [stdout]     |
[INFO] [stdout]     = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:562:47
[INFO] [stdout]     |
[INFO] [stdout] 484 | impl<B: Backend> CrossAttention<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 562 |         let attn_output = attn_weights.matmul(&v)?;
[INFO] [stdout]     |                                        ------ ^^ expected `&Tensor`, found `&Tensor<B>`
[INFO] [stdout]     |                                        |
[INFO] [stdout]     |                                        arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<Cpu>`
[INFO] [stdout]                found reference `&Tensor<B>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:139:12
[INFO] [stdout]     |
[INFO] [stdout] 139 |     pub fn matmul(&self, other: &Self) -> Result<Self> {
[INFO] [stdout]     |            ^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:567:42
[INFO] [stdout]     |
[INFO] [stdout] 484 | impl<B: Backend> CrossAttention<B> {
[INFO] [stdout]     |      - expected this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 567 |         let output = self.o_proj.forward(&attn_output)?;
[INFO] [stdout]     |                                  ------- ^^^^^^^^^^^^ expected `&Tensor<B>`, found `&Tensor`
[INFO] [stdout]     |                                  |
[INFO] [stdout]     |                                  arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<B>`
[INFO] [stdout]                found reference `&Tensor<Cpu>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:574:47
[INFO] [stdout]     |
[INFO] [stdout] 574 |         params.extend(prefix_params("q_proj", self.q_proj.parameters()));
[INFO] [stdout]     |                       -------------           ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:575:47
[INFO] [stdout]     |
[INFO] [stdout] 575 |         params.extend(prefix_params("k_proj", self.k_proj.parameters()));
[INFO] [stdout]     |                       -------------           ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:576:47
[INFO] [stdout]     |
[INFO] [stdout] 576 |         params.extend(prefix_params("v_proj", self.v_proj.parameters()));
[INFO] [stdout]     |                       -------------           ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:577:47
[INFO] [stdout]     |
[INFO] [stdout] 577 |         params.extend(prefix_params("o_proj", self.o_proj.parameters()));
[INFO] [stdout]     |                       -------------           ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try`
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:628:63
[INFO] [stdout]     |
[INFO] [stdout] 628 |             ActivationType::Gelu | ActivationType::GeluNew => h.gelu()?,
[INFO] [stdout]     |                                                               ^^^^^^^^^ the `?` operator cannot be applied to type `Tensor<B>`
[INFO] [stdout]     |
[INFO] [stdout]     = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try`
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:629:37
[INFO] [stdout]     |
[INFO] [stdout] 629 |             ActivationType::Relu => h.relu()?,
[INFO] [stdout]     |                                     ^^^^^^^^^ the `?` operator cannot be applied to type `Tensor<B>`
[INFO] [stdout]     |
[INFO] [stdout]     = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try`
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:630:37
[INFO] [stdout]     |
[INFO] [stdout] 630 |             ActivationType::Silu => h.silu()?,
[INFO] [stdout]     |                                     ^^^^^^^^^ the `?` operator cannot be applied to type `Tensor<B>`
[INFO] [stdout]     |
[INFO] [stdout]     = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try`
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:631:37
[INFO] [stdout]     |
[INFO] [stdout] 631 |             ActivationType::Tanh => h.tanh()?,
[INFO] [stdout]     |                                     ^^^^^^^^^ the `?` operator cannot be applied to type `Tensor<B>`
[INFO] [stdout]     |
[INFO] [stdout]     = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `mul` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:637:23
[INFO] [stdout]     |
[INFO] [stdout] 637 |                 h = h.mul(&gate)?;
[INFO] [stdout]     |                       ^^^ this is an associated function, not a method
[INFO] [stdout]     |
[INFO] [stdout]    --> /rustc/59807616e1fa2540724bfbac14d7976d7e4a3860/library/core/src/ops/arith.rs:339:7
[INFO] [stdout]     |
[INFO] [stdout]     = note: the method is available for `Tensor<B>` here
[INFO] [stdout]     |
[INFO] [stdout]     = note: found the following associated functions; to be used as methods, functions must have a `self` parameter
[INFO] [stdout] note: the candidate is defined in the trait `Backend`
[INFO] [stdout]    --> crates/myllm-core/src/backend/mod.rs:64:5
[INFO] [stdout]     |
[INFO] [stdout]  64 |     fn mul(a: &Self::Storage, b: &Self::Storage) -> Result<Self::Storage>;
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     = help: items from traits can only be used if the trait is in scope
[INFO] [stdout] help: use associated function syntax instead
[INFO] [stdout]     |
[INFO] [stdout] 637 -                 h = h.mul(&gate)?;
[INFO] [stdout] 637 +                 h = Tensor<B>::mul(&gate)?;
[INFO] [stdout]     |
[INFO] [stdout] help: trait `Mul` which provides `mul` is implemented but not in scope; perhaps you want to import it
[INFO] [stdout]     |
[INFO] [stdout]   9 + use std::ops::Mul;
[INFO] [stdout]     |
[INFO] [stdout] help: there is a method `matmul` with a similar name
[INFO] [stdout]     |
[INFO] [stdout] 637 |                 h = h.matmul(&gate)?;
[INFO] [stdout]     |                       +++
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:641:34
[INFO] [stdout]     |
[INFO] [stdout] 592 | impl<B: Backend> EncoderDecoderFFN<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 641 |         h = self.dropout.forward(&h)?;
[INFO] [stdout]     |                          ------- ^^ expected `&Tensor`, found `&Tensor<B>`
[INFO] [stdout]     |                          |
[INFO] [stdout]     |                          arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<Cpu>`
[INFO] [stdout]                found reference `&Tensor<B>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: `?` operator has incompatible types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:641:13
[INFO] [stdout]     |
[INFO] [stdout] 592 | impl<B: Backend> EncoderDecoderFFN<B> {
[INFO] [stdout]     |      - expected this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 641 |         h = self.dropout.forward(&h)?;
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^ expected `Tensor<B>`, found `Tensor`
[INFO] [stdout]     |
[INFO] [stdout]     = note: `?` operator cannot convert from `Tensor<Cpu>` to `Tensor<B>`
[INFO] [stdout]     = note: expected struct `Tensor<B>`
[INFO] [stdout]                found struct `Tensor<Cpu>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:648:43
[INFO] [stdout]     |
[INFO] [stdout] 648 |         params.extend(prefix_params("wi", self.wi.parameters()));
[INFO] [stdout]     |                       -------------       ^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:649:43
[INFO] [stdout]     |
[INFO] [stdout] 649 |         params.extend(prefix_params("wo", self.wo.parameters()));
[INFO] [stdout]     |                       -------------       ^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:652:49
[INFO] [stdout]     |
[INFO] [stdout] 652 |             params.extend(prefix_params("wi_1", wi_1.parameters()));
[INFO] [stdout]     |                           -------------         ^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                           |
[INFO] [stdout]     |                           arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:674:34
[INFO] [stdout]     |
[INFO] [stdout] 674 |             self_attention_norm: RMSNorm::new(config.d_model, config.layer_norm_eps),
[INFO] [stdout]     |                                  ^^^^^^^^^^^^ --------------  --------------------- unexpected argument #2 of type `f32`
[INFO] [stdout]     |                                               |
[INFO] [stdout]     |                                               expected `RMSNormConfig`, found `usize`
[INFO] [stdout]     |
[INFO] [stdout] note: associated function defined here
[INFO] [stdout]    --> crates/myllm-nn/src/norm.rs:185:12
[INFO] [stdout]     |
[INFO] [stdout] 185 |     pub fn new(config: RMSNormConfig) -> Self {
[INFO] [stdout]     |            ^^^
[INFO] [stdout] help: remove the extra argument
[INFO] [stdout]     |
[INFO] [stdout] 674 -             self_attention_norm: RMSNorm::new(config.d_model, config.layer_norm_eps),
[INFO] [stdout] 674 +             self_attention_norm: RMSNorm::new(/* RMSNormConfig */),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:676:23
[INFO] [stdout]     |
[INFO] [stdout] 676 |             ffn_norm: RMSNorm::new(config.d_model, config.layer_norm_eps),
[INFO] [stdout]     |                       ^^^^^^^^^^^^ --------------  --------------------- unexpected argument #2 of type `f32`
[INFO] [stdout]     |                                    |
[INFO] [stdout]     |                                    expected `RMSNormConfig`, found `usize`
[INFO] [stdout]     |
[INFO] [stdout] note: associated function defined here
[INFO] [stdout]    --> crates/myllm-nn/src/norm.rs:185:12
[INFO] [stdout]     |
[INFO] [stdout] 185 |     pub fn new(config: RMSNormConfig) -> Self {
[INFO] [stdout]     |            ^^^
[INFO] [stdout] help: remove the extra argument
[INFO] [stdout]     |
[INFO] [stdout] 676 -             ffn_norm: RMSNorm::new(config.d_model, config.layer_norm_eps),
[INFO] [stdout] 676 +             ffn_norm: RMSNorm::new(/* RMSNormConfig */),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:692:52
[INFO] [stdout]     |
[INFO] [stdout] 669 | impl<B: Backend> T5EncoderLayer<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 692 |             let attn_output = self.dropout.forward(&attn_output)?;
[INFO] [stdout]     |                                            ------- ^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor<B>`
[INFO] [stdout]     |                                            |
[INFO] [stdout]     |                                            arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<Cpu>`
[INFO] [stdout]                found reference `&Tensor<B>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `add` found for reference `&Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:693:47
[INFO] [stdout]     |
[INFO] [stdout] 693 |             let hidden_states = hidden_states.add(&attn_output)?;
[INFO] [stdout]     |                                               ^^^ this is an associated function, not a method
[INFO] [stdout]     |
[INFO] [stdout]     = note: found the following associated functions; to be used as methods, functions must have a `self` parameter
[INFO] [stdout] note: the candidate is defined in the trait `Backend`
[INFO] [stdout]    --> crates/myllm-core/src/backend/mod.rs:58:5
[INFO] [stdout]     |
[INFO] [stdout]  58 |     fn add(a: &Self::Storage, b: &Self::Storage) -> Result<Self::Storage>;
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     = help: items from traits can only be used if the trait is in scope
[INFO] [stdout] help: use associated function syntax instead
[INFO] [stdout]     |
[INFO] [stdout] 693 -             let hidden_states = hidden_states.add(&attn_output)?;
[INFO] [stdout] 693 +             let hidden_states = Tensor<B>::add(&attn_output)?;
[INFO] [stdout]     |
[INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it
[INFO] [stdout]     |
[INFO] [stdout]   9 + use std::ops::Add;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:697:51
[INFO] [stdout]     |
[INFO] [stdout] 669 | impl<B: Backend> T5EncoderLayer<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 697 |             let ffn_output = self.dropout.forward(&ffn_output)?;
[INFO] [stdout]     |                                           ------- ^^^^^^^^^^^ expected `&Tensor`, found `&Tensor<B>`
[INFO] [stdout]     |                                           |
[INFO] [stdout]     |                                           arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<Cpu>`
[INFO] [stdout]                found reference `&Tensor<B>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:703:52
[INFO] [stdout]     |
[INFO] [stdout] 669 | impl<B: Backend> T5EncoderLayer<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 703 |             let attn_output = self.dropout.forward(&attn_output)?;
[INFO] [stdout]     |                                            ------- ^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor<B>`
[INFO] [stdout]     |                                            |
[INFO] [stdout]     |                                            arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<Cpu>`
[INFO] [stdout]                found reference `&Tensor<B>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `add` found for reference `&Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:704:47
[INFO] [stdout]     |
[INFO] [stdout] 704 |             let hidden_states = hidden_states.add(&attn_output)?;
[INFO] [stdout]     |                                               ^^^ this is an associated function, not a method
[INFO] [stdout]     |
[INFO] [stdout]     = note: found the following associated functions; to be used as methods, functions must have a `self` parameter
[INFO] [stdout] note: the candidate is defined in the trait `Backend`
[INFO] [stdout]    --> crates/myllm-core/src/backend/mod.rs:58:5
[INFO] [stdout]     |
[INFO] [stdout]  58 |     fn add(a: &Self::Storage, b: &Self::Storage) -> Result<Self::Storage>;
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     = help: items from traits can only be used if the trait is in scope
[INFO] [stdout] help: use associated function syntax instead
[INFO] [stdout]     |
[INFO] [stdout] 704 -             let hidden_states = hidden_states.add(&attn_output)?;
[INFO] [stdout] 704 +             let hidden_states = Tensor<B>::add(&attn_output)?;
[INFO] [stdout]     |
[INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it
[INFO] [stdout]     |
[INFO] [stdout]   9 + use std::ops::Add;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:708:51
[INFO] [stdout]     |
[INFO] [stdout] 669 | impl<B: Backend> T5EncoderLayer<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 708 |             let ffn_output = self.dropout.forward(&ffn_output)?;
[INFO] [stdout]     |                                           ------- ^^^^^^^^^^^ expected `&Tensor`, found `&Tensor<B>`
[INFO] [stdout]     |                                           |
[INFO] [stdout]     |                                           arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<Cpu>`
[INFO] [stdout]                found reference `&Tensor<B>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `add` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:709:47
[INFO] [stdout]     |
[INFO] [stdout] 709 |             let hidden_states = hidden_states.add(&ffn_output)?;
[INFO] [stdout]     |                                               ^^^ this is an associated function, not a method
[INFO] [stdout]     |
[INFO] [stdout]    --> /rustc/59807616e1fa2540724bfbac14d7976d7e4a3860/library/core/src/ops/arith.rs:92:7
[INFO] [stdout]     |
[INFO] [stdout]     = note: the method is available for `Tensor<B>` here
[INFO] [stdout]     |
[INFO] [stdout]     = note: found the following associated functions; to be used as methods, functions must have a `self` parameter
[INFO] [stdout] note: the candidate is defined in the trait `Backend`
[INFO] [stdout]    --> crates/myllm-core/src/backend/mod.rs:58:5
[INFO] [stdout]     |
[INFO] [stdout]  58 |     fn add(a: &Self::Storage, b: &Self::Storage) -> Result<Self::Storage>;
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     = help: items from traits can only be used if the trait is in scope
[INFO] [stdout] help: use associated function syntax instead
[INFO] [stdout]     |
[INFO] [stdout] 709 -             let hidden_states = hidden_states.add(&ffn_output)?;
[INFO] [stdout] 709 +             let hidden_states = Tensor<B>::add(&ffn_output)?;
[INFO] [stdout]     |
[INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it
[INFO] [stdout]     |
[INFO] [stdout]   9 + use std::ops::Add;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:720:13
[INFO] [stdout]     |
[INFO] [stdout] 718 |         params.extend(prefix_params(
[INFO] [stdout]     |                       ------------- arguments to this function are incorrect
[INFO] [stdout] 719 |             "self_attention_norm",
[INFO] [stdout] 720 |             self.self_attention_norm.parameters(),
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:723:49
[INFO] [stdout]     |
[INFO] [stdout] 723 |         params.extend(prefix_params("ffn_norm", self.ffn_norm.parameters()));
[INFO] [stdout]     |                       -------------             ^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:745:34
[INFO] [stdout]     |
[INFO] [stdout] 745 |             self_attention_norm: RMSNorm::new(config.d_model, config.layer_norm_eps),
[INFO] [stdout]     |                                  ^^^^^^^^^^^^ --------------  --------------------- unexpected argument #2 of type `f32`
[INFO] [stdout]     |                                               |
[INFO] [stdout]     |                                               expected `RMSNormConfig`, found `usize`
[INFO] [stdout]     |
[INFO] [stdout] note: associated function defined here
[INFO] [stdout]    --> crates/myllm-nn/src/norm.rs:185:12
[INFO] [stdout]     |
[INFO] [stdout] 185 |     pub fn new(config: RMSNormConfig) -> Self {
[INFO] [stdout]     |            ^^^
[INFO] [stdout] help: remove the extra argument
[INFO] [stdout]     |
[INFO] [stdout] 745 -             self_attention_norm: RMSNorm::new(config.d_model, config.layer_norm_eps),
[INFO] [stdout] 745 +             self_attention_norm: RMSNorm::new(/* RMSNormConfig */),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:747:35
[INFO] [stdout]     |
[INFO] [stdout] 747 |             cross_attention_norm: RMSNorm::new(config.d_model, config.layer_norm_eps),
[INFO] [stdout]     |                                   ^^^^^^^^^^^^ --------------  --------------------- unexpected argument #2 of type `f32`
[INFO] [stdout]     |                                                |
[INFO] [stdout]     |                                                expected `RMSNormConfig`, found `usize`
[INFO] [stdout]     |
[INFO] [stdout] note: associated function defined here
[INFO] [stdout]    --> crates/myllm-nn/src/norm.rs:185:12
[INFO] [stdout]     |
[INFO] [stdout] 185 |     pub fn new(config: RMSNormConfig) -> Self {
[INFO] [stdout]     |            ^^^
[INFO] [stdout] help: remove the extra argument
[INFO] [stdout]     |
[INFO] [stdout] 747 -             cross_attention_norm: RMSNorm::new(config.d_model, config.layer_norm_eps),
[INFO] [stdout] 747 +             cross_attention_norm: RMSNorm::new(/* RMSNormConfig */),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:749:23
[INFO] [stdout]     |
[INFO] [stdout] 749 |             ffn_norm: RMSNorm::new(config.d_model, config.layer_norm_eps),
[INFO] [stdout]     |                       ^^^^^^^^^^^^ --------------  --------------------- unexpected argument #2 of type `f32`
[INFO] [stdout]     |                                    |
[INFO] [stdout]     |                                    expected `RMSNormConfig`, found `usize`
[INFO] [stdout]     |
[INFO] [stdout] note: associated function defined here
[INFO] [stdout]    --> crates/myllm-nn/src/norm.rs:185:12
[INFO] [stdout]     |
[INFO] [stdout] 185 |     pub fn new(config: RMSNormConfig) -> Self {
[INFO] [stdout]     |            ^^^
[INFO] [stdout] help: remove the extra argument
[INFO] [stdout]     |
[INFO] [stdout] 749 -             ffn_norm: RMSNorm::new(config.d_model, config.layer_norm_eps),
[INFO] [stdout] 749 +             ffn_norm: RMSNorm::new(/* RMSNormConfig */),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:768:52
[INFO] [stdout]     |
[INFO] [stdout] 740 | impl<B: Backend> T5DecoderLayer<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 768 |             let attn_output = self.dropout.forward(&attn_output)?;
[INFO] [stdout]     |                                            ------- ^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor<B>`
[INFO] [stdout]     |                                            |
[INFO] [stdout]     |                                            arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<Cpu>`
[INFO] [stdout]                found reference `&Tensor<B>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `add` found for reference `&Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:769:35
[INFO] [stdout]     |
[INFO] [stdout] 769 |             let h = hidden_states.add(&attn_output)?;
[INFO] [stdout]     |                                   ^^^ this is an associated function, not a method
[INFO] [stdout]     |
[INFO] [stdout]     = note: found the following associated functions; to be used as methods, functions must have a `self` parameter
[INFO] [stdout] note: the candidate is defined in the trait `Backend`
[INFO] [stdout]    --> crates/myllm-core/src/backend/mod.rs:58:5
[INFO] [stdout]     |
[INFO] [stdout]  58 |     fn add(a: &Self::Storage, b: &Self::Storage) -> Result<Self::Storage>;
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     = help: items from traits can only be used if the trait is in scope
[INFO] [stdout] help: use associated function syntax instead
[INFO] [stdout]     |
[INFO] [stdout] 769 -             let h = hidden_states.add(&attn_output)?;
[INFO] [stdout] 769 +             let h = Tensor<B>::add(&attn_output)?;
[INFO] [stdout]     |
[INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it
[INFO] [stdout]     |
[INFO] [stdout]   9 + use std::ops::Add;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:775:52
[INFO] [stdout]     |
[INFO] [stdout] 740 | impl<B: Backend> T5DecoderLayer<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 775 |             let attn_output = self.dropout.forward(&attn_output)?;
[INFO] [stdout]     |                                            ------- ^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor<B>`
[INFO] [stdout]     |                                            |
[INFO] [stdout]     |                                            arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<Cpu>`
[INFO] [stdout]                found reference `&Tensor<B>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `add` found for reference `&Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:776:35
[INFO] [stdout]     |
[INFO] [stdout] 776 |             let h = hidden_states.add(&attn_output)?;
[INFO] [stdout]     |                                   ^^^ this is an associated function, not a method
[INFO] [stdout]     |
[INFO] [stdout]     = note: found the following associated functions; to be used as methods, functions must have a `self` parameter
[INFO] [stdout] note: the candidate is defined in the trait `Backend`
[INFO] [stdout]    --> crates/myllm-core/src/backend/mod.rs:58:5
[INFO] [stdout]     |
[INFO] [stdout]  58 |     fn add(a: &Self::Storage, b: &Self::Storage) -> Result<Self::Storage>;
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     = help: items from traits can only be used if the trait is in scope
[INFO] [stdout] help: use associated function syntax instead
[INFO] [stdout]     |
[INFO] [stdout] 776 -             let h = hidden_states.add(&attn_output)?;
[INFO] [stdout] 776 +             let h = Tensor<B>::add(&attn_output)?;
[INFO] [stdout]     |
[INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it
[INFO] [stdout]     |
[INFO] [stdout]   9 + use std::ops::Add;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:789:53
[INFO] [stdout]     |
[INFO] [stdout] 740 | impl<B: Backend> T5DecoderLayer<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 789 |             let cross_output = self.dropout.forward(&cross_output)?;
[INFO] [stdout]     |                                             ------- ^^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor<B>`
[INFO] [stdout]     |                                             |
[INFO] [stdout]     |                                             arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<Cpu>`
[INFO] [stdout]                found reference `&Tensor<B>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `add` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:790:35
[INFO] [stdout]     |
[INFO] [stdout] 790 |             let h = hidden_states.add(&cross_output)?;
[INFO] [stdout]     |                                   ^^^ this is an associated function, not a method
[INFO] [stdout]     |
[INFO] [stdout]    --> /rustc/59807616e1fa2540724bfbac14d7976d7e4a3860/library/core/src/ops/arith.rs:92:7
[INFO] [stdout]     |
[INFO] [stdout]     = note: the method is available for `Tensor<B>` here
[INFO] [stdout]     |
[INFO] [stdout]     = note: found the following associated functions; to be used as methods, functions must have a `self` parameter
[INFO] [stdout] note: the candidate is defined in the trait `Backend`
[INFO] [stdout]    --> crates/myllm-core/src/backend/mod.rs:58:5
[INFO] [stdout]     |
[INFO] [stdout]  58 |     fn add(a: &Self::Storage, b: &Self::Storage) -> Result<Self::Storage>;
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     = help: items from traits can only be used if the trait is in scope
[INFO] [stdout] help: use associated function syntax instead
[INFO] [stdout]     |
[INFO] [stdout] 790 -             let h = hidden_states.add(&cross_output)?;
[INFO] [stdout] 790 +             let h = Tensor<B>::add(&cross_output)?;
[INFO] [stdout]     |
[INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it
[INFO] [stdout]     |
[INFO] [stdout]   9 + use std::ops::Add;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:799:53
[INFO] [stdout]     |
[INFO] [stdout] 740 | impl<B: Backend> T5DecoderLayer<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 799 |             let cross_output = self.dropout.forward(&cross_output)?;
[INFO] [stdout]     |                                             ------- ^^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor<B>`
[INFO] [stdout]     |                                             |
[INFO] [stdout]     |                                             arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<Cpu>`
[INFO] [stdout]                found reference `&Tensor<B>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `add` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:800:35
[INFO] [stdout]     |
[INFO] [stdout] 800 |             let h = hidden_states.add(&cross_output)?;
[INFO] [stdout]     |                                   ^^^ this is an associated function, not a method
[INFO] [stdout]     |
[INFO] [stdout]    --> /rustc/59807616e1fa2540724bfbac14d7976d7e4a3860/library/core/src/ops/arith.rs:92:7
[INFO] [stdout]     |
[INFO] [stdout]     = note: the method is available for `Tensor<B>` here
[INFO] [stdout]     |
[INFO] [stdout]     = note: found the following associated functions; to be used as methods, functions must have a `self` parameter
[INFO] [stdout] note: the candidate is defined in the trait `Backend`
[INFO] [stdout]    --> crates/myllm-core/src/backend/mod.rs:58:5
[INFO] [stdout]     |
[INFO] [stdout]  58 |     fn add(a: &Self::Storage, b: &Self::Storage) -> Result<Self::Storage>;
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     = help: items from traits can only be used if the trait is in scope
[INFO] [stdout] help: use associated function syntax instead
[INFO] [stdout]     |
[INFO] [stdout] 800 -             let h = hidden_states.add(&cross_output)?;
[INFO] [stdout] 800 +             let h = Tensor<B>::add(&cross_output)?;
[INFO] [stdout]     |
[INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it
[INFO] [stdout]     |
[INFO] [stdout]   9 + use std::ops::Add;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:808:51
[INFO] [stdout]     |
[INFO] [stdout] 740 | impl<B: Backend> T5DecoderLayer<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 808 |             let ffn_output = self.dropout.forward(&ffn_output)?;
[INFO] [stdout]     |                                           ------- ^^^^^^^^^^^ expected `&Tensor`, found `&Tensor<B>`
[INFO] [stdout]     |                                           |
[INFO] [stdout]     |                                           arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<Cpu>`
[INFO] [stdout]                found reference `&Tensor<B>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `add` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:809:27
[INFO] [stdout]     |
[INFO] [stdout] 809 |             hidden_states.add(&ffn_output)?
[INFO] [stdout]     |                           ^^^ this is an associated function, not a method
[INFO] [stdout]     |
[INFO] [stdout]    --> /rustc/59807616e1fa2540724bfbac14d7976d7e4a3860/library/core/src/ops/arith.rs:92:7
[INFO] [stdout]     |
[INFO] [stdout]     = note: the method is available for `Tensor<B>` here
[INFO] [stdout]     |
[INFO] [stdout]     = note: found the following associated functions; to be used as methods, functions must have a `self` parameter
[INFO] [stdout] note: the candidate is defined in the trait `Backend`
[INFO] [stdout]    --> crates/myllm-core/src/backend/mod.rs:58:5
[INFO] [stdout]     |
[INFO] [stdout]  58 |     fn add(a: &Self::Storage, b: &Self::Storage) -> Result<Self::Storage>;
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     = help: items from traits can only be used if the trait is in scope
[INFO] [stdout] help: use associated function syntax instead
[INFO] [stdout]     |
[INFO] [stdout] 809 -             hidden_states.add(&ffn_output)?
[INFO] [stdout] 809 +             Tensor<B>::add(&ffn_output)?
[INFO] [stdout]     |
[INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it
[INFO] [stdout]     |
[INFO] [stdout]   9 + use std::ops::Add;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:812:51
[INFO] [stdout]     |
[INFO] [stdout] 740 | impl<B: Backend> T5DecoderLayer<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 812 |             let ffn_output = self.dropout.forward(&ffn_output)?;
[INFO] [stdout]     |                                           ------- ^^^^^^^^^^^ expected `&Tensor`, found `&Tensor<B>`
[INFO] [stdout]     |                                           |
[INFO] [stdout]     |                                           arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<Cpu>`
[INFO] [stdout]                found reference `&Tensor<B>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `add` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:813:35
[INFO] [stdout]     |
[INFO] [stdout] 813 |             let h = hidden_states.add(&ffn_output)?;
[INFO] [stdout]     |                                   ^^^ this is an associated function, not a method
[INFO] [stdout]     |
[INFO] [stdout]    --> /rustc/59807616e1fa2540724bfbac14d7976d7e4a3860/library/core/src/ops/arith.rs:92:7
[INFO] [stdout]     |
[INFO] [stdout]     = note: the method is available for `Tensor<B>` here
[INFO] [stdout]     |
[INFO] [stdout]     = note: found the following associated functions; to be used as methods, functions must have a `self` parameter
[INFO] [stdout] note: the candidate is defined in the trait `Backend`
[INFO] [stdout]    --> crates/myllm-core/src/backend/mod.rs:58:5
[INFO] [stdout]     |
[INFO] [stdout]  58 |     fn add(a: &Self::Storage, b: &Self::Storage) -> Result<Self::Storage>;
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     = help: items from traits can only be used if the trait is in scope
[INFO] [stdout] help: use associated function syntax instead
[INFO] [stdout]     |
[INFO] [stdout] 813 -             let h = hidden_states.add(&ffn_output)?;
[INFO] [stdout] 813 +             let h = Tensor<B>::add(&ffn_output)?;
[INFO] [stdout]     |
[INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it
[INFO] [stdout]     |
[INFO] [stdout]   9 + use std::ops::Add;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:831:13
[INFO] [stdout]     |
[INFO] [stdout] 829 |         params.extend(prefix_params(
[INFO] [stdout]     |                       ------------- arguments to this function are incorrect
[INFO] [stdout] 830 |             "self_attention_norm",
[INFO] [stdout] 831 |             self.self_attention_norm.parameters(),
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:836:13
[INFO] [stdout]     |
[INFO] [stdout] 834 |         params.extend(prefix_params(
[INFO] [stdout]     |                       ------------- arguments to this function are incorrect
[INFO] [stdout] 835 |             "cross_attention_norm",
[INFO] [stdout] 836 |             self.cross_attention_norm.parameters(),
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:839:49
[INFO] [stdout]     |
[INFO] [stdout] 839 |         params.extend(prefix_params("ffn_norm", self.ffn_norm.parameters()));
[INFO] [stdout]     |                       -------------             ^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:899:27
[INFO] [stdout]     |
[INFO] [stdout] 899 |             embed_tokens: Embedding::new(config.vocab_size, config.d_model),
[INFO] [stdout]     |                           ^^^^^^^^^^^^^^ -----------------  -------------- unexpected argument #2 of type `usize`
[INFO] [stdout]     |                                          |
[INFO] [stdout]     |                                          expected `EmbeddingConfig`, found `usize`
[INFO] [stdout]     |
[INFO] [stdout] note: associated function defined here
[INFO] [stdout]    --> crates/myllm-nn/src/embedding.rs:35:12
[INFO] [stdout]     |
[INFO] [stdout]  35 |     pub fn new(config: EmbeddingConfig) -> Self {
[INFO] [stdout]     |            ^^^
[INFO] [stdout] help: remove the extra argument
[INFO] [stdout]     |
[INFO] [stdout] 899 -             embed_tokens: Embedding::new(config.vocab_size, config.d_model),
[INFO] [stdout] 899 +             embed_tokens: Embedding::new(/* EmbeddingConfig */),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:901:25
[INFO] [stdout]     |
[INFO] [stdout] 901 |             final_norm: RMSNorm::new(config.d_model, config.layer_norm_eps),
[INFO] [stdout]     |                         ^^^^^^^^^^^^ --------------  --------------------- unexpected argument #2 of type `f32`
[INFO] [stdout]     |                                      |
[INFO] [stdout]     |                                      expected `RMSNormConfig`, found `usize`
[INFO] [stdout]     |
[INFO] [stdout] note: associated function defined here
[INFO] [stdout]    --> crates/myllm-nn/src/norm.rs:185:12
[INFO] [stdout]     |
[INFO] [stdout] 185 |     pub fn new(config: RMSNormConfig) -> Self {
[INFO] [stdout]     |            ^^^
[INFO] [stdout] help: remove the extra argument
[INFO] [stdout]     |
[INFO] [stdout] 901 -             final_norm: RMSNorm::new(config.d_model, config.layer_norm_eps),
[INFO] [stdout] 901 +             final_norm: RMSNorm::new(/* RMSNormConfig */),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:916:54
[INFO] [stdout]     |
[INFO] [stdout] 885 | impl<B: Backend> T5Encoder<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 916 |         let mut hidden_states = self.dropout.forward(&hidden_states)?;
[INFO] [stdout]     |                                              ------- ^^^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor<B>`
[INFO] [stdout]     |                                              |
[INFO] [stdout]     |                                              arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<Cpu>`
[INFO] [stdout]                found reference `&Tensor<B>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:925:43
[INFO] [stdout]     |
[INFO] [stdout] 885 | impl<B: Backend> T5Encoder<B> {
[INFO] [stdout]     |      - expected this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 925 |             hidden_states = layer.forward(&hidden_states, attention_mask, position_bias.as_ref())?;
[INFO] [stdout]     |                                   ------- ^^^^^^^^^^^^^^ expected `&Tensor<B>`, found `&Tensor`
[INFO] [stdout]     |                                   |
[INFO] [stdout]     |                                   arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<B>`
[INFO] [stdout]                found reference `&Tensor<Cpu>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:683:12
[INFO] [stdout]     |
[INFO] [stdout] 683 |     pub fn forward(
[INFO] [stdout]     |            ^^^^^^^
[INFO] [stdout] 684 |         &self,
[INFO] [stdout] 685 |         hidden_states: &Tensor<B>,
[INFO] [stdout]     |         -------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: `?` operator has incompatible types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:925:29
[INFO] [stdout]     |
[INFO] [stdout] 885 | impl<B: Backend> T5Encoder<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 925 |             hidden_states = layer.forward(&hidden_states, attention_mask, position_bias.as_ref())?;
[INFO] [stdout]     |                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `Tensor`, found `Tensor<B>`
[INFO] [stdout]     |
[INFO] [stdout]     = note: `?` operator cannot convert from `Tensor<B>` to `Tensor<Cpu>`
[INFO] [stdout]     = note: expected struct `Tensor<Cpu>`
[INFO] [stdout]                found struct `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:928:33
[INFO] [stdout]     |
[INFO] [stdout] 885 | impl<B: Backend> T5Encoder<B> {
[INFO] [stdout]     |      - expected this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 928 |         self.final_norm.forward(&hidden_states)
[INFO] [stdout]     |                         ------- ^^^^^^^^^^^^^^ expected `&Tensor<B>`, found `&Tensor`
[INFO] [stdout]     |                         |
[INFO] [stdout]     |                         arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<B>`
[INFO] [stdout]                found reference `&Tensor<Cpu>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:934:53
[INFO] [stdout]     |
[INFO] [stdout] 934 |         params.extend(prefix_params("embed_tokens", self.embed_tokens.parameters()));
[INFO] [stdout]     |                       -------------                 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:940:51
[INFO] [stdout]     |
[INFO] [stdout] 940 |         params.extend(prefix_params("final_norm", self.final_norm.parameters()));
[INFO] [stdout]     |                       -------------               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:973:27
[INFO] [stdout]     |
[INFO] [stdout] 973 |             embed_tokens: Embedding::new(config.vocab_size, config.d_model),
[INFO] [stdout]     |                           ^^^^^^^^^^^^^^ -----------------  -------------- unexpected argument #2 of type `usize`
[INFO] [stdout]     |                                          |
[INFO] [stdout]     |                                          expected `EmbeddingConfig`, found `usize`
[INFO] [stdout]     |
[INFO] [stdout] note: associated function defined here
[INFO] [stdout]    --> crates/myllm-nn/src/embedding.rs:35:12
[INFO] [stdout]     |
[INFO] [stdout]  35 |     pub fn new(config: EmbeddingConfig) -> Self {
[INFO] [stdout]     |            ^^^
[INFO] [stdout] help: remove the extra argument
[INFO] [stdout]     |
[INFO] [stdout] 973 -             embed_tokens: Embedding::new(config.vocab_size, config.d_model),
[INFO] [stdout] 973 +             embed_tokens: Embedding::new(/* EmbeddingConfig */),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:975:25
[INFO] [stdout]     |
[INFO] [stdout] 975 |             final_norm: RMSNorm::new(config.d_model, config.layer_norm_eps),
[INFO] [stdout]     |                         ^^^^^^^^^^^^ --------------  --------------------- unexpected argument #2 of type `f32`
[INFO] [stdout]     |                                      |
[INFO] [stdout]     |                                      expected `RMSNormConfig`, found `usize`
[INFO] [stdout]     |
[INFO] [stdout] note: associated function defined here
[INFO] [stdout]    --> crates/myllm-nn/src/norm.rs:185:12
[INFO] [stdout]     |
[INFO] [stdout] 185 |     pub fn new(config: RMSNormConfig) -> Self {
[INFO] [stdout]     |            ^^^
[INFO] [stdout] help: remove the extra argument
[INFO] [stdout]     |
[INFO] [stdout] 975 -             final_norm: RMSNorm::new(config.d_model, config.layer_norm_eps),
[INFO] [stdout] 975 +             final_norm: RMSNorm::new(/* RMSNormConfig */),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/encoder_decoder.rs:997:54
[INFO] [stdout]     |
[INFO] [stdout] 959 | impl<B: Backend> T5Decoder<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 997 |         let mut hidden_states = self.dropout.forward(&hidden_states)?;
[INFO] [stdout]     |                                              ------- ^^^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor<B>`
[INFO] [stdout]     |                                              |
[INFO] [stdout]     |                                              arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<Cpu>`
[INFO] [stdout]                found reference `&Tensor<B>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]     --> crates/myllm-model/src/encoder_decoder.rs:1011:17
[INFO] [stdout]      |
[INFO] [stdout]  959 | impl<B: Backend> T5Decoder<B> {
[INFO] [stdout]      |      - expected this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 1010 |             let (output, layer_new_cache) = layer.forward(
[INFO] [stdout]      |                                                   ------- arguments to this method are incorrect
[INFO] [stdout] 1011 |                 &hidden_states,
[INFO] [stdout]      |                 ^^^^^^^^^^^^^^ expected `&Tensor<B>`, found `&Tensor`
[INFO] [stdout]      |
[INFO] [stdout]      = note: expected reference `&Tensor<B>`
[INFO] [stdout]                 found reference `&Tensor<Cpu>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]     --> crates/myllm-model/src/encoder_decoder.rs:756:12
[INFO] [stdout]      |
[INFO] [stdout]  756 |     pub fn forward(
[INFO] [stdout]      |            ^^^^^^^
[INFO] [stdout]  757 |         &self,
[INFO] [stdout]  758 |         hidden_states: &Tensor<B>,
[INFO] [stdout]      |         -------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]     --> crates/myllm-model/src/encoder_decoder.rs:1019:29
[INFO] [stdout]      |
[INFO] [stdout]  959 | impl<B: Backend> T5Decoder<B> {
[INFO] [stdout]      |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout]  997 |         let mut hidden_states = self.dropout.forward(&hidden_states)?;
[INFO] [stdout]      |                                 ------------------------------------- expected due to this value
[INFO] [stdout] ...
[INFO] [stdout] 1019 |             hidden_states = output;
[INFO] [stdout]      |                             ^^^^^^ expected `Tensor`, found `Tensor<B>`
[INFO] [stdout]      |
[INFO] [stdout]      = note: expected struct `Tensor<Cpu>`
[INFO] [stdout]                 found struct `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]     --> crates/myllm-model/src/encoder_decoder.rs:1023:53
[INFO] [stdout]      |
[INFO] [stdout]  959 | impl<B: Backend> T5Decoder<B> {
[INFO] [stdout]      |      - expected this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 1023 |         let hidden_states = self.final_norm.forward(&hidden_states)?;
[INFO] [stdout]      |                                             ------- ^^^^^^^^^^^^^^ expected `&Tensor<B>`, found `&Tensor`
[INFO] [stdout]      |                                             |
[INFO] [stdout]      |                                             arguments to this method are incorrect
[INFO] [stdout]      |
[INFO] [stdout]      = note: expected reference `&Tensor<B>`
[INFO] [stdout]                 found reference `&Tensor<Cpu>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]     --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]      |
[INFO] [stdout]    9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]      |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]     --> crates/myllm-model/src/encoder_decoder.rs:1035:53
[INFO] [stdout]      |
[INFO] [stdout] 1035 |         params.extend(prefix_params("embed_tokens", self.embed_tokens.parameters()));
[INFO] [stdout]      |                       -------------                 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]      |                       |
[INFO] [stdout]      |                       arguments to this function are incorrect
[INFO] [stdout]      |
[INFO] [stdout]      = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                 found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]     --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]      |
[INFO] [stdout]  808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]      |        ^^^^^^^^^^^^^
[INFO] [stdout]  809 |     prefix: &str,
[INFO] [stdout]  810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]      |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]     --> crates/myllm-model/src/encoder_decoder.rs:1041:51
[INFO] [stdout]      |
[INFO] [stdout] 1041 |         params.extend(prefix_params("final_norm", self.final_norm.parameters()));
[INFO] [stdout]      |                       -------------               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]      |                       |
[INFO] [stdout]      |                       arguments to this function are incorrect
[INFO] [stdout]      |
[INFO] [stdout]      = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                 found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]     --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]      |
[INFO] [stdout]  808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]      |        ^^^^^^^^^^^^^
[INFO] [stdout]  809 |     prefix: &str,
[INFO] [stdout]  810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]      |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]     --> crates/myllm-model/src/encoder_decoder.rs:1138:48
[INFO] [stdout]      |
[INFO] [stdout] 1138 |         params.extend(prefix_params("lm_head", self.lm_head.parameters()));
[INFO] [stdout]      |                       -------------            ^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]      |                       |
[INFO] [stdout]      |                       arguments to this function are incorrect
[INFO] [stdout]      |
[INFO] [stdout]      = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                 found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]     --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]      |
[INFO] [stdout]  808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]      |        ^^^^^^^^^^^^^
[INFO] [stdout]  809 |     prefix: &str,
[INFO] [stdout]  810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]      |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no function or associated item named `full` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]     --> crates/myllm-model/src/encoder_decoder.rs:1173:37
[INFO] [stdout]      |
[INFO] [stdout] 1173 |         let decoder_start = Tensor::full([batch_size, 1], self.model.config.decoder_start_token_id as f32)?;
[INFO] [stdout]      |                                     ^^^^ function or associated item not found in `Tensor<_>`
[INFO] [stdout]      |
[INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions:
[INFO] [stdout]       Tensor::<B>::from_storage
[INFO] [stdout]       Tensor::<B>::from_slice
[INFO] [stdout]       Tensor::<B>::zeros
[INFO] [stdout]       Tensor::<B>::ones
[INFO] [stdout]       and 7 others
[INFO] [stdout]     --> crates/myllm-core/src/tensor.rs:20:5
[INFO] [stdout]      |
[INFO] [stdout]   20 |     pub fn from_storage(storage: B::Storage) -> Self {
[INFO] [stdout]      |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]   29 |     pub fn from_slice(data: &[f32], shape: impl Into<Shape>) -> Self {
[INFO] [stdout]      |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]   35 |     pub fn zeros(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]      |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]   40 |     pub fn ones(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]      |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0277]: `DecoderCache<B>` doesn't implement `Debug`
[INFO] [stdout]     --> crates/myllm-model/src/encoder_decoder.rs:1198:5
[INFO] [stdout]      |
[INFO] [stdout] 1191 | #[derive(Debug)]
[INFO] [stdout]      |          ----- in this derive macro expansion
[INFO] [stdout] ...
[INFO] [stdout] 1198 |     pub cache: DecoderCache<B>,
[INFO] [stdout]      |     ^^^^^^^^^^^^^^^^^^^^^^^^^^ unsatisfied trait bound
[INFO] [stdout]      |
[INFO] [stdout] help: the trait `Debug` is not implemented for `DecoderCache<B>`
[INFO] [stdout]     --> crates/myllm-model/src/encoder_decoder.rs:859:1
[INFO] [stdout]      |
[INFO] [stdout]  859 | pub struct DecoderCache<B: Backend = Cpu> {
[INFO] [stdout]      | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]      = note: add `#[derive(Debug)]` to `DecoderCache<B>` or manually `impl Debug for DecoderCache<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no function or associated item named `from_vec` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]     --> crates/myllm-model/src/encoder_decoder.rs:1215:13
[INFO] [stdout]      |
[INFO] [stdout] 1215 |     Tensor::from_vec(mask, [query_len, key_len])
[INFO] [stdout]      |             ^^^^^^^^ function or associated item not found in `Tensor<_>`
[INFO] [stdout]      |
[INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions:
[INFO] [stdout]       Tensor::<B>::from_storage
[INFO] [stdout]       Tensor::<B>::from_slice
[INFO] [stdout]       Tensor::<B>::zeros
[INFO] [stdout]       Tensor::<B>::ones
[INFO] [stdout]       and 7 others
[INFO] [stdout]     --> crates/myllm-core/src/tensor.rs:20:5
[INFO] [stdout]      |
[INFO] [stdout]   20 |     pub fn from_storage(storage: B::Storage) -> Self {
[INFO] [stdout]      |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]   29 |     pub fn from_slice(data: &[f32], shape: impl Into<Shape>) -> Self {
[INFO] [stdout]      |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]   35 |     pub fn zeros(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]      |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]   40 |     pub fn ones(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]      |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] help: there is an associated function `from` with a similar name
[INFO] [stdout]     --> /rustc/59807616e1fa2540724bfbac14d7976d7e4a3860/library/core/src/convert/mod.rs:592:4
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:182:47
[INFO] [stdout]     |
[INFO] [stdout] 131 | impl<B: Backend> DualEncoder<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 182 |         let embeddings = self.dropout.forward(&embeddings)?;
[INFO] [stdout]     |                                       ------- ^^^^^^^^^^^ expected `&Tensor`, found `&Tensor<B>`
[INFO] [stdout]     |                                       |
[INFO] [stdout]     |                                       arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<Cpu>`
[INFO] [stdout]                found reference `&Tensor<B>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:185:26
[INFO] [stdout]     |
[INFO] [stdout] 131 | impl<B: Backend> DualEncoder<B> {
[INFO] [stdout]     |      - expected this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 185 |             proj.forward(&embeddings)?
[INFO] [stdout]     |                  ------- ^^^^^^^^^^^ expected `&Tensor<B>`, found `&Tensor`
[INFO] [stdout]     |                  |
[INFO] [stdout]     |                  arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<B>`
[INFO] [stdout]                found reference `&Tensor<Cpu>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: `if` and `else` have incompatible types
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:187:13
[INFO] [stdout]     |
[INFO] [stdout] 131 |   impl<B: Backend> DualEncoder<B> {
[INFO] [stdout]     |        - expected this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 184 |           let embeddings = if let Some(ref proj) = self.query_projection {
[INFO] [stdout]     |  __________________________-
[INFO] [stdout] 185 | |             proj.forward(&embeddings)?
[INFO] [stdout]     | |             -------------------------- expected because of this
[INFO] [stdout] 186 | |         } else {
[INFO] [stdout] 187 | |             embeddings
[INFO] [stdout]     | |             ^^^^^^^^^^ expected `Tensor<B>`, found `Tensor`
[INFO] [stdout] 188 | |         };
[INFO] [stdout]     | |_________- `if` and `else` have incompatible types
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `Tensor<B>`
[INFO] [stdout]                found struct `Tensor<Cpu>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:205:47
[INFO] [stdout]     |
[INFO] [stdout] 131 | impl<B: Backend> DualEncoder<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 205 |         let embeddings = self.dropout.forward(&embeddings)?;
[INFO] [stdout]     |                                       ------- ^^^^^^^^^^^ expected `&Tensor`, found `&Tensor<B>`
[INFO] [stdout]     |                                       |
[INFO] [stdout]     |                                       arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<Cpu>`
[INFO] [stdout]                found reference `&Tensor<B>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:214:26
[INFO] [stdout]     |
[INFO] [stdout] 131 | impl<B: Backend> DualEncoder<B> {
[INFO] [stdout]     |      - expected this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 214 |             proj.forward(&embeddings)?
[INFO] [stdout]     |                  ------- ^^^^^^^^^^^ expected `&Tensor<B>`, found `&Tensor`
[INFO] [stdout]     |                  |
[INFO] [stdout]     |                  arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<B>`
[INFO] [stdout]                found reference `&Tensor<Cpu>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: `if` and `else` have incompatible types
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:216:13
[INFO] [stdout]     |
[INFO] [stdout] 131 |   impl<B: Backend> DualEncoder<B> {
[INFO] [stdout]     |        - expected this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 213 |           let embeddings = if let Some(proj) = projection {
[INFO] [stdout]     |  __________________________-
[INFO] [stdout] 214 | |             proj.forward(&embeddings)?
[INFO] [stdout]     | |             -------------------------- expected because of this
[INFO] [stdout] 215 | |         } else {
[INFO] [stdout] 216 | |             embeddings
[INFO] [stdout]     | |             ^^^^^^^^^^ expected `Tensor<B>`, found `Tensor`
[INFO] [stdout] 217 | |         };
[INFO] [stdout]     | |_________- `if` and `else` have incompatible types
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `Tensor<B>`
[INFO] [stdout]                found struct `Tensor<Cpu>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this method takes 0 arguments but 2 arguments were supplied
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:232:62
[INFO] [stdout]     |
[INFO] [stdout] 232 |         let scores = query_embeddings.matmul(&doc_embeddings.transpose(0, 1)?)?;
[INFO] [stdout]     |                                                              ^^^^^^^^^ -  - unexpected argument #2 of type `{integer}`
[INFO] [stdout]     |                                                                        |
[INFO] [stdout]     |                                                                        unexpected argument #1 of type `{integer}`
[INFO] [stdout]     |
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:149:12
[INFO] [stdout]     |
[INFO] [stdout] 149 |     pub fn transpose(&self) -> Result<Self> {
[INFO] [stdout]     |            ^^^^^^^^^
[INFO] [stdout] help: remove the extra arguments
[INFO] [stdout]     |
[INFO] [stdout] 232 -         let scores = query_embeddings.matmul(&doc_embeddings.transpose(0, 1)?)?;
[INFO] [stdout] 232 +         let scores = query_embeddings.matmul(&doc_embeddings.transpose()?)?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `div_scalar` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:233:16
[INFO] [stdout]     |
[INFO] [stdout] 233 |         scores.div_scalar(self.config.temperature)
[INFO] [stdout]     |                ^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: there is a method `add_scalar` with a similar name
[INFO] [stdout]     |
[INFO] [stdout] 233 -         scores.div_scalar(self.config.temperature)
[INFO] [stdout] 233 +         scores.add_scalar(self.config.temperature)
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:270:61
[INFO] [stdout]     |
[INFO] [stdout] 270 |             params.extend(prefix_params("query_projection", proj.parameters()));
[INFO] [stdout]     |                           -------------                     ^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                           |
[INFO] [stdout]     |                           arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:274:59
[INFO] [stdout]     |
[INFO] [stdout] 274 |             params.extend(prefix_params("doc_projection", proj.parameters()));
[INFO] [stdout]     |                           -------------                   ^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                           |
[INFO] [stdout]     |                           arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:357:47
[INFO] [stdout]     |
[INFO] [stdout] 304 | impl<B: Backend> ColBERT<B> {
[INFO] [stdout]     |      - found this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 357 |         let embeddings = self.dropout.forward(&embeddings)?;
[INFO] [stdout]     |                                       ------- ^^^^^^^^^^^ expected `&Tensor`, found `&Tensor<B>`
[INFO] [stdout]     |                                       |
[INFO] [stdout]     |                                       arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<Cpu>`
[INFO] [stdout]                found reference `&Tensor<B>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `unsqueeze` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:369:60
[INFO] [stdout]     |
[INFO] [stdout] 369 |             let mask_expanded = mask.slice(1, 0, max_len)?.unsqueeze(-1)?;
[INFO] [stdout]     |                                                            ^^^^^^^^^ method not found in `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `mul` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:370:41
[INFO] [stdout]     |
[INFO] [stdout] 370 |             let embeddings = embeddings.mul(&mask_expanded)?;
[INFO] [stdout]     |                                         ^^^ this is an associated function, not a method
[INFO] [stdout]     |
[INFO] [stdout]    --> /rustc/59807616e1fa2540724bfbac14d7976d7e4a3860/library/core/src/ops/arith.rs:339:7
[INFO] [stdout]     |
[INFO] [stdout]     = note: the method is available for `Tensor` here
[INFO] [stdout]     |
[INFO] [stdout]     = note: found the following associated functions; to be used as methods, functions must have a `self` parameter
[INFO] [stdout] note: the candidate is defined in the trait `Backend`
[INFO] [stdout]    --> crates/myllm-core/src/backend/mod.rs:64:5
[INFO] [stdout]     |
[INFO] [stdout]  64 |     fn mul(a: &Self::Storage, b: &Self::Storage) -> Result<Self::Storage>;
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     = help: items from traits can only be used if the trait is in scope
[INFO] [stdout] help: use associated function syntax instead
[INFO] [stdout]     |
[INFO] [stdout] 370 -             let embeddings = embeddings.mul(&mask_expanded)?;
[INFO] [stdout] 370 +             let embeddings = Tensor::mul(&mask_expanded)?;
[INFO] [stdout]     |
[INFO] [stdout] help: trait `Mul` which provides `mul` is implemented but not in scope; perhaps you want to import it
[INFO] [stdout]     |
[INFO] [stdout]   9 + use std::ops::Mul;
[INFO] [stdout]     |
[INFO] [stdout] help: there is a method `matmul` with a similar name
[INFO] [stdout]     |
[INFO] [stdout] 370 |             let embeddings = embeddings.matmul(&mask_expanded)?;
[INFO] [stdout]     |                                         +++
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:378:35
[INFO] [stdout]     |
[INFO] [stdout] 304 | impl<B: Backend> ColBERT<B> {
[INFO] [stdout]     |      - expected this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 378 |             normalize_l2_last_dim(&embeddings)
[INFO] [stdout]     |             --------------------- ^^^^^^^^^^^ expected `&Tensor<B>`, found `&Tensor`
[INFO] [stdout]     |             |
[INFO] [stdout]     |             arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<B>`
[INFO] [stdout]                found reference `&Tensor<Cpu>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:785:4
[INFO] [stdout]     |
[INFO] [stdout] 785 | fn normalize_l2_last_dim<B: Backend>(tensor: &Tensor<B>) -> Result<Tensor<B>> {
[INFO] [stdout]     |    ^^^^^^^^^^^^^^^^^^^^^             ------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:380:16
[INFO] [stdout]     |
[INFO] [stdout] 304 | impl<B: Backend> ColBERT<B> {
[INFO] [stdout]     |      - expected this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 380 |             Ok(embeddings)
[INFO] [stdout]     |             -- ^^^^^^^^^^ expected `Tensor<B>`, found `Tensor`
[INFO] [stdout]     |             |
[INFO] [stdout]     |             arguments to this enum variant are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `Tensor<B>`
[INFO] [stdout]                found struct `Tensor<Cpu>`
[INFO] [stdout] help: the type constructed contains `Tensor` due to the type of the argument passed
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:380:13
[INFO] [stdout]     |
[INFO] [stdout] 380 |             Ok(embeddings)
[INFO] [stdout]     |             ^^^----------^
[INFO] [stdout]     |                |
[INFO] [stdout]     |                this argument influences the type of `Ok`
[INFO] [stdout] note: tuple variant defined here
[INFO] [stdout]    --> /rustc/59807616e1fa2540724bfbac14d7976d7e4a3860/library/core/src/result.rs:561:4
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this method takes 0 arguments but 2 arguments were supplied
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:409:35
[INFO] [stdout]     |
[INFO] [stdout] 409 |             let sim = q.matmul(&d.transpose(0, 1)?)?;
[INFO] [stdout]     |                                   ^^^^^^^^^ -  - unexpected argument #2 of type `{integer}`
[INFO] [stdout]     |                                             |
[INFO] [stdout]     |                                             unexpected argument #1 of type `{integer}`
[INFO] [stdout]     |
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:149:12
[INFO] [stdout]     |
[INFO] [stdout] 149 |     pub fn transpose(&self) -> Result<Self> {
[INFO] [stdout]     |            ^^^^^^^^^
[INFO] [stdout] help: remove the extra arguments
[INFO] [stdout]     |
[INFO] [stdout] 409 -             let sim = q.matmul(&d.transpose(0, 1)?)?;
[INFO] [stdout] 409 +             let sim = q.matmul(&d.transpose()?)?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no function or associated item named `stack` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:415:17
[INFO] [stdout]     |
[INFO] [stdout] 415 |         Tensor::stack(&scores.iter().collect::<Vec<_>>(), 0)
[INFO] [stdout]     |                 ^^^^^ function or associated item not found in `Tensor<_>`
[INFO] [stdout]     |
[INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions:
[INFO] [stdout]       Tensor::<B>::from_storage
[INFO] [stdout]       Tensor::<B>::from_slice
[INFO] [stdout]       Tensor::<B>::zeros
[INFO] [stdout]       Tensor::<B>::ones
[INFO] [stdout]       and 7 others
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:20:5
[INFO] [stdout]     |
[INFO] [stdout]  20 |     pub fn from_storage(storage: B::Storage) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  29 |     pub fn from_slice(data: &[f32], shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  35 |     pub fn zeros(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  40 |     pub fn ones(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:446:47
[INFO] [stdout]     |
[INFO] [stdout] 446 |         params.extend(prefix_params("linear", self.linear.parameters()));
[INFO] [stdout]     |                       -------------           ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:508:46
[INFO] [stdout]     |
[INFO] [stdout] 474 | impl<B: Backend> CrossEncoder<B> {
[INFO] [stdout]     |      - expected this type parameter
[INFO] [stdout] ...
[INFO] [stdout] 508 |         let scores = self.classifier.forward(&pooled)?;
[INFO] [stdout]     |                                      ------- ^^^^^^^ expected `&Tensor<B>`, found `&Tensor`
[INFO] [stdout]     |                                      |
[INFO] [stdout]     |                                      arguments to this method are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected reference `&Tensor<B>`
[INFO] [stdout]                found reference `&Tensor<Cpu>`
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-nn/src/layer.rs:9:8
[INFO] [stdout]     |
[INFO] [stdout]   9 |     fn forward(&self, input: &Tensor<B>) -> Result<Tensor<B>>;
[INFO] [stdout]     |        ^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `squeeze` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:510:16
[INFO] [stdout]     |
[INFO] [stdout] 510 |         scores.squeeze(-1)
[INFO] [stdout]     |                ^^^^^^^ method not found in `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:522:51
[INFO] [stdout]     |
[INFO] [stdout] 522 |         params.extend(prefix_params("classifier", self.classifier.parameters()));
[INFO] [stdout]     |                       -------------               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap<String, Tensor<_>>`, found `Vec<&Tensor<B>>`
[INFO] [stdout]     |                       |
[INFO] [stdout]     |                       arguments to this function are incorrect
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected struct `HashMap<String, Tensor<_>>`
[INFO] [stdout]                found struct `Vec<&Tensor<B>>`
[INFO] [stdout] note: function defined here
[INFO] [stdout]    --> crates/myllm-model/src/encoder.rs:808:8
[INFO] [stdout]     |
[INFO] [stdout] 808 | pub fn prefix_params<B: Backend>(
[INFO] [stdout]     |        ^^^^^^^^^^^^^
[INFO] [stdout] 809 |     prefix: &str,
[INFO] [stdout] 810 |     params: HashMap<String, Tensor<B>>,
[INFO] [stdout]     |     ----------------------------------
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `div_scalar` found for reference `&Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:537:36
[INFO] [stdout]     |
[INFO] [stdout] 537 |         let scaled_scores = scores.div_scalar(temperature)?;
[INFO] [stdout]     |                                    ^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: there is a method `add_scalar` with a similar name
[INFO] [stdout]     |
[INFO] [stdout] 537 -         let scaled_scores = scores.div_scalar(temperature)?;
[INFO] [stdout] 537 +         let scaled_scores = scores.add_scalar(temperature)?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no function or associated item named `from_vec` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:552:17
[INFO] [stdout]     |
[INFO] [stdout] 552 |         Tensor::from_vec(vec![loss_sum / batch_size as f32], [1])
[INFO] [stdout]     |                 ^^^^^^^^ function or associated item not found in `Tensor<_>`
[INFO] [stdout]     |
[INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions:
[INFO] [stdout]       Tensor::<B>::from_storage
[INFO] [stdout]       Tensor::<B>::from_slice
[INFO] [stdout]       Tensor::<B>::zeros
[INFO] [stdout]       Tensor::<B>::ones
[INFO] [stdout]       and 7 others
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:20:5
[INFO] [stdout]     |
[INFO] [stdout]  20 |     pub fn from_storage(storage: B::Storage) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  29 |     pub fn from_slice(data: &[f32], shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  35 |     pub fn zeros(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  40 |     pub fn ones(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] help: there is an associated function `from` with a similar name
[INFO] [stdout]    --> /rustc/59807616e1fa2540724bfbac14d7976d7e4a3860/library/core/src/convert/mod.rs:592:4
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this method takes 0 arguments but 2 arguments were supplied
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:564:57
[INFO] [stdout]     |
[INFO] [stdout] 564 |         let pos_scores = query_emb.matmul(&positive_emb.transpose(0, 1)?)?;
[INFO] [stdout]     |                                                         ^^^^^^^^^ -  - unexpected argument #2 of type `{integer}`
[INFO] [stdout]     |                                                                   |
[INFO] [stdout]     |                                                                   unexpected argument #1 of type `{integer}`
[INFO] [stdout]     |
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:149:12
[INFO] [stdout]     |
[INFO] [stdout] 149 |     pub fn transpose(&self) -> Result<Self> {
[INFO] [stdout]     |            ^^^^^^^^^
[INFO] [stdout] help: remove the extra arguments
[INFO] [stdout]     |
[INFO] [stdout] 564 -         let pos_scores = query_emb.matmul(&positive_emb.transpose(0, 1)?)?;
[INFO] [stdout] 564 +         let pos_scores = query_emb.matmul(&positive_emb.transpose()?)?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `div_scalar` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:565:37
[INFO] [stdout]     |
[INFO] [stdout] 565 |         let pos_scores = pos_scores.div_scalar(temperature)?;
[INFO] [stdout]     |                                     ^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: there is a method `add_scalar` with a similar name
[INFO] [stdout]     |
[INFO] [stdout] 565 -         let pos_scores = pos_scores.div_scalar(temperature)?;
[INFO] [stdout] 565 +         let pos_scores = pos_scores.add_scalar(temperature)?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this method takes 0 arguments but 2 arguments were supplied
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:568:56
[INFO] [stdout]     |
[INFO] [stdout] 568 |             let neg_scores = query_emb.matmul(&neg_emb.transpose(0, 1)?)?;
[INFO] [stdout]     |                                                        ^^^^^^^^^ -  - unexpected argument #2 of type `{integer}`
[INFO] [stdout]     |                                                                  |
[INFO] [stdout]     |                                                                  unexpected argument #1 of type `{integer}`
[INFO] [stdout]     |
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:149:12
[INFO] [stdout]     |
[INFO] [stdout] 149 |     pub fn transpose(&self) -> Result<Self> {
[INFO] [stdout]     |            ^^^^^^^^^
[INFO] [stdout] help: remove the extra arguments
[INFO] [stdout]     |
[INFO] [stdout] 568 -             let neg_scores = query_emb.matmul(&neg_emb.transpose(0, 1)?)?;
[INFO] [stdout] 568 +             let neg_scores = query_emb.matmul(&neg_emb.transpose()?)?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `div_scalar` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:569:41
[INFO] [stdout]     |
[INFO] [stdout] 569 |             let neg_scores = neg_scores.div_scalar(temperature)?;
[INFO] [stdout]     |                                         ^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: there is a method `add_scalar` with a similar name
[INFO] [stdout]     |
[INFO] [stdout] 569 -             let neg_scores = neg_scores.div_scalar(temperature)?;
[INFO] [stdout] 569 +             let neg_scores = neg_scores.add_scalar(temperature)?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no function or associated item named `from_vec` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:590:17
[INFO] [stdout]     |
[INFO] [stdout] 590 |         Tensor::from_vec(vec![loss_sum / batch_size as f32], [1])
[INFO] [stdout]     |                 ^^^^^^^^ function or associated item not found in `Tensor<_>`
[INFO] [stdout]     |
[INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions:
[INFO] [stdout]       Tensor::<B>::from_storage
[INFO] [stdout]       Tensor::<B>::from_slice
[INFO] [stdout]       Tensor::<B>::zeros
[INFO] [stdout]       Tensor::<B>::ones
[INFO] [stdout]       and 7 others
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:20:5
[INFO] [stdout]     |
[INFO] [stdout]  20 |     pub fn from_storage(storage: B::Storage) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  29 |     pub fn from_slice(data: &[f32], shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  35 |     pub fn zeros(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  40 |     pub fn ones(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] help: there is an associated function `from` with a similar name
[INFO] [stdout]    --> /rustc/59807616e1fa2540724bfbac14d7976d7e4a3860/library/core/src/convert/mod.rs:592:4
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `sub` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:603:30
[INFO] [stdout]     |
[INFO] [stdout] 603 |         let loss = (pos_dist.sub(&neg_dist)?.add_scalar(margin)?).relu()?;
[INFO] [stdout]     |                              ^^^ this is an associated function, not a method
[INFO] [stdout]     |
[INFO] [stdout]    --> /rustc/59807616e1fa2540724bfbac14d7976d7e4a3860/library/core/src/ops/arith.rs:205:7
[INFO] [stdout]     |
[INFO] [stdout]     = note: the method is available for `Tensor<B>` here
[INFO] [stdout]     |
[INFO] [stdout]     = note: found the following associated functions; to be used as methods, functions must have a `self` parameter
[INFO] [stdout] note: the candidate is defined in the trait `Backend`
[INFO] [stdout]    --> crates/myllm-core/src/backend/mod.rs:61:5
[INFO] [stdout]     |
[INFO] [stdout]  61 |     fn sub(a: &Self::Storage, b: &Self::Storage) -> Result<Self::Storage>;
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     = help: items from traits can only be used if the trait is in scope
[INFO] [stdout] help: there is a method `sum` with a similar name, but with different arguments
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:219:5
[INFO] [stdout]     |
[INFO] [stdout] 219 |     pub fn sum(&self) -> f32 {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] help: use associated function syntax instead
[INFO] [stdout]     |
[INFO] [stdout] 603 -         let loss = (pos_dist.sub(&neg_dist)?.add_scalar(margin)?).relu()?;
[INFO] [stdout] 603 +         let loss = (Tensor<B>::sub(&neg_dist)?.add_scalar(margin)?).relu()?;
[INFO] [stdout]     |
[INFO] [stdout] help: trait `Sub` which provides `sub` is implemented but not in scope; perhaps you want to import it
[INFO] [stdout]     |
[INFO] [stdout]   9 + use std::ops::Sub;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this method takes 0 arguments but 2 arguments were supplied
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:615:54
[INFO] [stdout]     |
[INFO] [stdout] 615 |         let pos_sim = query_emb.matmul(&positive_emb.transpose(0, 1)?)?;
[INFO] [stdout]     |                                                      ^^^^^^^^^ -  - unexpected argument #2 of type `{integer}`
[INFO] [stdout]     |                                                                |
[INFO] [stdout]     |                                                                unexpected argument #1 of type `{integer}`
[INFO] [stdout]     |
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:149:12
[INFO] [stdout]     |
[INFO] [stdout] 149 |     pub fn transpose(&self) -> Result<Self> {
[INFO] [stdout]     |            ^^^^^^^^^
[INFO] [stdout] help: remove the extra arguments
[INFO] [stdout]     |
[INFO] [stdout] 615 -         let pos_sim = query_emb.matmul(&positive_emb.transpose(0, 1)?)?;
[INFO] [stdout] 615 +         let pos_sim = query_emb.matmul(&positive_emb.transpose()?)?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this method takes 0 arguments but 2 arguments were supplied
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:616:54
[INFO] [stdout]     |
[INFO] [stdout] 616 |         let neg_sim = query_emb.matmul(&negative_emb.transpose(0, 1)?)?;
[INFO] [stdout]     |                                                      ^^^^^^^^^ -  - unexpected argument #2 of type `{integer}`
[INFO] [stdout]     |                                                                |
[INFO] [stdout]     |                                                                unexpected argument #1 of type `{integer}`
[INFO] [stdout]     |
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:149:12
[INFO] [stdout]     |
[INFO] [stdout] 149 |     pub fn transpose(&self) -> Result<Self> {
[INFO] [stdout]     |            ^^^^^^^^^
[INFO] [stdout] help: remove the extra arguments
[INFO] [stdout]     |
[INFO] [stdout] 616 -         let neg_sim = query_emb.matmul(&negative_emb.transpose(0, 1)?)?;
[INFO] [stdout] 616 +         let neg_sim = query_emb.matmul(&negative_emb.transpose()?)?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `sub_scalar` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:621:33
[INFO] [stdout]     |
[INFO] [stdout] 621 |         let delta_pos = pos_sim.sub_scalar(opt_pos)?.neg()?;
[INFO] [stdout]     |                                 ^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: there is a method `add_scalar` with a similar name
[INFO] [stdout]     |
[INFO] [stdout] 621 -         let delta_pos = pos_sim.sub_scalar(opt_pos)?.neg()?;
[INFO] [stdout] 621 +         let delta_pos = pos_sim.add_scalar(opt_pos)?.neg()?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `sub_scalar` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:622:33
[INFO] [stdout]     |
[INFO] [stdout] 622 |         let delta_neg = neg_sim.sub_scalar(opt_neg)?;
[INFO] [stdout]     |                                 ^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: there is a method `add_scalar` with a similar name
[INFO] [stdout]     |
[INFO] [stdout] 622 -         let delta_neg = neg_sim.sub_scalar(opt_neg)?;
[INFO] [stdout] 622 +         let delta_neg = neg_sim.add_scalar(opt_neg)?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `sub_scalar` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:628:27
[INFO] [stdout]     |
[INFO] [stdout] 628 |             .mul(&pos_sim.sub_scalar(1.0 - margin)?)?
[INFO] [stdout]     |                           ^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: there is a method `add_scalar` with a similar name
[INFO] [stdout]     |
[INFO] [stdout] 628 -             .mul(&pos_sim.sub_scalar(1.0 - margin)?)?
[INFO] [stdout] 628 +             .mul(&pos_sim.add_scalar(1.0 - margin)?)?
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try`
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:631:19
[INFO] [stdout]     |
[INFO] [stdout] 631 |             .mul(&neg_sim.add_scalar(margin)?)?
[INFO] [stdout]     |                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^ the `?` operator cannot be applied to type `Tensor<B>`
[INFO] [stdout]     |
[INFO] [stdout]     = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `to_vec` found for reference `&Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:677:34
[INFO] [stdout]     |
[INFO] [stdout] 677 |         let scores_data = scores.to_vec()?;
[INFO] [stdout]     |                                  ^^^^^^ method not found in `&Tensor<B>`
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this method takes 0 arguments but 2 arguments were supplied
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:723:62
[INFO] [stdout]     |
[INFO] [stdout] 723 |         let scores = query_embeddings.matmul(&doc_embeddings.transpose(0, 1)?)?;
[INFO] [stdout]     |                                                              ^^^^^^^^^ -  - unexpected argument #2 of type `{integer}`
[INFO] [stdout]     |                                                                        |
[INFO] [stdout]     |                                                                        unexpected argument #1 of type `{integer}`
[INFO] [stdout]     |
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:149:12
[INFO] [stdout]     |
[INFO] [stdout] 149 |     pub fn transpose(&self) -> Result<Self> {
[INFO] [stdout]     |            ^^^^^^^^^
[INFO] [stdout] help: remove the extra arguments
[INFO] [stdout]     |
[INFO] [stdout] 723 -         let scores = query_embeddings.matmul(&doc_embeddings.transpose(0, 1)?)?;
[INFO] [stdout] 723 +         let scores = query_embeddings.matmul(&doc_embeddings.transpose()?)?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no function or associated item named `from_vec` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:727:30
[INFO] [stdout]     |
[INFO] [stdout] 727 |         let labels = Tensor::from_vec(labels, [batch_size])?;
[INFO] [stdout]     |                              ^^^^^^^^ function or associated item not found in `Tensor<_>`
[INFO] [stdout]     |
[INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions:
[INFO] [stdout]       Tensor::<B>::from_storage
[INFO] [stdout]       Tensor::<B>::from_slice
[INFO] [stdout]       Tensor::<B>::zeros
[INFO] [stdout]       Tensor::<B>::ones
[INFO] [stdout]       and 7 others
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:20:5
[INFO] [stdout]     |
[INFO] [stdout]  20 |     pub fn from_storage(storage: B::Storage) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  29 |     pub fn from_slice(data: &[f32], shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  35 |     pub fn zeros(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  40 |     pub fn ones(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] help: there is an associated function `from` with a similar name
[INFO] [stdout]    --> /rustc/59807616e1fa2540724bfbac14d7976d7e4a3860/library/core/src/convert/mod.rs:592:4
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this method takes 0 arguments but 2 arguments were supplied
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:744:71
[INFO] [stdout]     |
[INFO] [stdout] 744 |         let pos_scores = query_embeddings.matmul(&positive_embeddings.transpose(0, 1)?)?;
[INFO] [stdout]     |                                                                       ^^^^^^^^^ -  - unexpected argument #2 of type `{integer}`
[INFO] [stdout]     |                                                                                 |
[INFO] [stdout]     |                                                                                 unexpected argument #1 of type `{integer}`
[INFO] [stdout]     |
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:149:12
[INFO] [stdout]     |
[INFO] [stdout] 149 |     pub fn transpose(&self) -> Result<Self> {
[INFO] [stdout]     |            ^^^^^^^^^
[INFO] [stdout] help: remove the extra arguments
[INFO] [stdout]     |
[INFO] [stdout] 744 -         let pos_scores = query_embeddings.matmul(&positive_embeddings.transpose(0, 1)?)?;
[INFO] [stdout] 744 +         let pos_scores = query_embeddings.matmul(&positive_embeddings.transpose()?)?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this method takes 0 arguments but 2 arguments were supplied
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:745:71
[INFO] [stdout]     |
[INFO] [stdout] 745 |         let neg_scores = query_embeddings.matmul(&negative_embeddings.transpose(0, 1)?)?;
[INFO] [stdout]     |                                                                       ^^^^^^^^^ -  - unexpected argument #2 of type `{integer}`
[INFO] [stdout]     |                                                                                 |
[INFO] [stdout]     |                                                                                 unexpected argument #1 of type `{integer}`
[INFO] [stdout]     |
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:149:12
[INFO] [stdout]     |
[INFO] [stdout] 149 |     pub fn transpose(&self) -> Result<Self> {
[INFO] [stdout]     |            ^^^^^^^^^
[INFO] [stdout] help: remove the extra arguments
[INFO] [stdout]     |
[INFO] [stdout] 745 -         let neg_scores = query_embeddings.matmul(&negative_embeddings.transpose(0, 1)?)?;
[INFO] [stdout] 745 +         let neg_scores = query_embeddings.matmul(&negative_embeddings.transpose()?)?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no function or associated item named `from_vec` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:750:30
[INFO] [stdout]     |
[INFO] [stdout] 750 |         let labels = Tensor::from_vec(labels, [batch_size])?;
[INFO] [stdout]     |                              ^^^^^^^^ function or associated item not found in `Tensor<_>`
[INFO] [stdout]     |
[INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions:
[INFO] [stdout]       Tensor::<B>::from_storage
[INFO] [stdout]       Tensor::<B>::from_slice
[INFO] [stdout]       Tensor::<B>::zeros
[INFO] [stdout]       Tensor::<B>::ones
[INFO] [stdout]       and 7 others
[INFO] [stdout]    --> crates/myllm-core/src/tensor.rs:20:5
[INFO] [stdout]     |
[INFO] [stdout]  20 |     pub fn from_storage(storage: B::Storage) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  29 |     pub fn from_slice(data: &[f32], shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  35 |     pub fn zeros(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] ...
[INFO] [stdout]  40 |     pub fn ones(shape: impl Into<Shape>) -> Self {
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] help: there is an associated function `from` with a similar name
[INFO] [stdout]    --> /rustc/59807616e1fa2540724bfbac14d7976d7e4a3860/library/core/src/convert/mod.rs:592:4
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `pow_scalar` found for reference `&Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:779:23
[INFO] [stdout]     |
[INFO] [stdout] 779 |     let norm = tensor.pow_scalar(2.0)?.sum_dim(-1, true)?.sqrt()?;
[INFO] [stdout]     |                       ^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: there is a method `add_scalar` with a similar name
[INFO] [stdout]     |
[INFO] [stdout] 779 -     let norm = tensor.pow_scalar(2.0)?.sum_dim(-1, true)?.sqrt()?;
[INFO] [stdout] 779 +     let norm = tensor.add_scalar(2.0)?.sum_dim(-1, true)?.sqrt()?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `div` found for reference `&Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:781:12
[INFO] [stdout]     |
[INFO] [stdout] 781 |     tensor.div(&norm)
[INFO] [stdout]     |            ^^^ this is an associated function, not a method
[INFO] [stdout]     |
[INFO] [stdout]     = note: found the following associated functions; to be used as methods, functions must have a `self` parameter
[INFO] [stdout] note: the candidate is defined in the trait `Backend`
[INFO] [stdout]    --> crates/myllm-core/src/backend/mod.rs:67:5
[INFO] [stdout]     |
[INFO] [stdout]  67 |     fn div(a: &Self::Storage, b: &Self::Storage) -> Result<Self::Storage>;
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     = help: items from traits can only be used if the trait is in scope
[INFO] [stdout] help: use associated function syntax instead
[INFO] [stdout]     |
[INFO] [stdout] 781 -     tensor.div(&norm)
[INFO] [stdout] 781 +     Tensor<B>::div(&norm)
[INFO] [stdout]     |
[INFO] [stdout] help: trait `Div` which provides `div` is implemented but not in scope; perhaps you want to import it
[INFO] [stdout]     |
[INFO] [stdout]   9 + use std::ops::Div;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `pow_scalar` found for reference `&Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:786:23
[INFO] [stdout]     |
[INFO] [stdout] 786 |     let norm = tensor.pow_scalar(2.0)?.sum_dim(-1, true)?.sqrt()?;
[INFO] [stdout]     |                       ^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: there is a method `add_scalar` with a similar name
[INFO] [stdout]     |
[INFO] [stdout] 786 -     let norm = tensor.pow_scalar(2.0)?.sum_dim(-1, true)?.sqrt()?;
[INFO] [stdout] 786 +     let norm = tensor.add_scalar(2.0)?.sum_dim(-1, true)?.sqrt()?;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `div` found for reference `&Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:788:12
[INFO] [stdout]     |
[INFO] [stdout] 788 |     tensor.div(&norm)
[INFO] [stdout]     |            ^^^ this is an associated function, not a method
[INFO] [stdout]     |
[INFO] [stdout]     = note: found the following associated functions; to be used as methods, functions must have a `self` parameter
[INFO] [stdout] note: the candidate is defined in the trait `Backend`
[INFO] [stdout]    --> crates/myllm-core/src/backend/mod.rs:67:5
[INFO] [stdout]     |
[INFO] [stdout]  67 |     fn div(a: &Self::Storage, b: &Self::Storage) -> Result<Self::Storage>;
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     = help: items from traits can only be used if the trait is in scope
[INFO] [stdout] help: use associated function syntax instead
[INFO] [stdout]     |
[INFO] [stdout] 788 -     tensor.div(&norm)
[INFO] [stdout] 788 +     Tensor<B>::div(&norm)
[INFO] [stdout]     |
[INFO] [stdout] help: trait `Div` which provides `div` is implemented but not in scope; perhaps you want to import it
[INFO] [stdout]     |
[INFO] [stdout]   9 + use std::ops::Div;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `mul` found for struct `Tensor<B>` in the current scope
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:795:29
[INFO] [stdout]     |
[INFO] [stdout] 795 |     let similarity = a_norm.mul(&b_norm)?.sum_dim(-1, false)?;
[INFO] [stdout]     |                             ^^^ this is an associated function, not a method
[INFO] [stdout]     |
[INFO] [stdout]    --> /rustc/59807616e1fa2540724bfbac14d7976d7e4a3860/library/core/src/ops/arith.rs:339:7
[INFO] [stdout]     |
[INFO] [stdout]     = note: the method is available for `Tensor<B>` here
[INFO] [stdout]     |
[INFO] [stdout]     = note: found the following associated functions; to be used as methods, functions must have a `self` parameter
[INFO] [stdout] note: the candidate is defined in the trait `Backend`
[INFO] [stdout]    --> crates/myllm-core/src/backend/mod.rs:64:5
[INFO] [stdout]     |
[INFO] [stdout]  64 |     fn mul(a: &Self::Storage, b: &Self::Storage) -> Result<Self::Storage>;
[INFO] [stdout]     |     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     = help: items from traits can only be used if the trait is in scope
[INFO] [stdout] help: use associated function syntax instead
[INFO] [stdout]     |
[INFO] [stdout] 795 -     let similarity = a_norm.mul(&b_norm)?.sum_dim(-1, false)?;
[INFO] [stdout] 795 +     let similarity = Tensor<B>::mul(&b_norm)?.sum_dim(-1, false)?;
[INFO] [stdout]     |
[INFO] [stdout] help: trait `Mul` which provides `mul` is implemented but not in scope; perhaps you want to import it
[INFO] [stdout]     |
[INFO] [stdout]   9 + use std::ops::Mul;
[INFO] [stdout]     |
[INFO] [stdout] help: there is a method `matmul` with a similar name
[INFO] [stdout]     |
[INFO] [stdout] 795 |     let similarity = a_norm.matmul(&b_norm)?.sum_dim(-1, false)?;
[INFO] [stdout]     |                             +++
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stderr]    Compiling toml v0.8.23
[INFO] [stdout] warning: unused variable: `ratio`
[INFO] [stdout]    --> crates/myllm-model/src/context_compression.rs:319:13
[INFO] [stdout]     |
[INFO] [stdout] 319 |         let ratio = self.config.compression_ratio;
[INFO] [stdout]     |             ^^^^^ help: if this is intentional, prefix it with an underscore: `_ratio`
[INFO] [stdout]     |
[INFO] [stdout]     = note: `#[warn(unused_variables)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0382]: use of moved value: `config`
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:171:35
[INFO] [stdout]     |
[INFO] [stdout] 133 |     pub fn new(config: RetrievalConfig) -> Self {
[INFO] [stdout]     |                ------ move occurs because `config` has type `RetrievalConfig`, which does not implement the `Copy` trait
[INFO] [stdout] ...
[INFO] [stdout] 166 |             config,
[INFO] [stdout]     |             ------ value moved here
[INFO] [stdout] ...
[INFO] [stdout] 171 |             dropout: Dropout::new(config.dropout),
[INFO] [stdout]     |                                   ^^^^^^^^^^^^^^ value used here after move
[INFO] [stdout]     |
[INFO] [stdout] help: consider cloning the value if the performance cost is acceptable
[INFO] [stdout]     |
[INFO] [stdout] 166 |             config: config.clone(),
[INFO] [stdout]     |                   ++++++++++++++++
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0382]: use of moved value: `config`
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:320:35
[INFO] [stdout]     |
[INFO] [stdout] 306 |     pub fn new(config: RetrievalConfig) -> Self {
[INFO] [stdout]     |                ------ move occurs because `config` has type `RetrievalConfig`, which does not implement the `Copy` trait
[INFO] [stdout] ...
[INFO] [stdout] 317 |             config,
[INFO] [stdout]     |             ------ value moved here
[INFO] [stdout] ...
[INFO] [stdout] 320 |             dropout: Dropout::new(config.dropout),
[INFO] [stdout]     |                                   ^^^^^^^^^^^^^^ value used here after move
[INFO] [stdout]     |
[INFO] [stdout] help: consider cloning the value if the performance cost is acceptable
[INFO] [stdout]     |
[INFO] [stdout] 317 |             config: config.clone(),
[INFO] [stdout]     |                   ++++++++++++++++
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0382]: use of moved value: `config`
[INFO] [stdout]    --> crates/myllm-model/src/retrieval.rs:490:35
[INFO] [stdout]     |
[INFO] [stdout] 476 |     pub fn new(config: RetrievalConfig) -> Self {
[INFO] [stdout]     |                ------ move occurs because `config` has type `RetrievalConfig`, which does not implement the `Copy` trait
[INFO] [stdout] ...
[INFO] [stdout] 487 |             config,
[INFO] [stdout]     |             ------ value moved here
[INFO] [stdout] ...
[INFO] [stdout] 490 |             dropout: Dropout::new(config.dropout),
[INFO] [stdout]     |                                   ^^^^^^^^^^^^^^ value used here after move
[INFO] [stdout]     |
[INFO] [stdout] help: consider cloning the value if the performance cost is acceptable
[INFO] [stdout]     |
[INFO] [stdout] 487 |             config: config.clone(),
[INFO] [stdout]     |                   ++++++++++++++++
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] Some errors have detailed explanations: E0061, E0277, E0308, E0382, E0433, E0599, E0600.
[INFO] [stdout] 
[INFO] [stdout] For more information about an error, try `rustc --explain E0061`.
[INFO] [stdout] 
[INFO] [stderr] error: could not compile `myllm-model` (lib) due to 229 previous errors; 4 warnings emitted
[INFO] [stderr] warning: build failed, waiting for other jobs to finish...
[INFO] running `Command { std: "docker" "inspect" "3b8a792b6f56861e5eeeb785ed1a4504b056c622947088d978a17140fc70698a", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "3b8a792b6f56861e5eeeb785ed1a4504b056c622947088d978a17140fc70698a", kill_on_drop: false }`
[INFO] [stdout] 3b8a792b6f56861e5eeeb785ed1a4504b056c622947088d978a17140fc70698a
