[INFO] cloning repository https://github.com/comfortablynumb/myllm [INFO] running `Command { std: "git" "-c" "credential.helper=" "-c" "credential.helper=/workspace/cargo-home/bin/git-credential-null" "clone" "--bare" "https://github.com/comfortablynumb/myllm" "/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2Fcomfortablynumb%2Fmyllm", kill_on_drop: false }` [INFO] [stderr] Cloning into bare repository '/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2Fcomfortablynumb%2Fmyllm'... [INFO] running `Command { std: "git" "rev-parse" "HEAD", kill_on_drop: false }` [INFO] [stdout] 3f4031598d5eba236544b8bfcd136cf3c4163221 [INFO] testing comfortablynumb/myllm against beta-2026-04-21 for beta-1.96-1 [INFO] running `Command { std: "git" "clone" "/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2Fcomfortablynumb%2Fmyllm" "/workspace/builds/worker-6-tc2/source", kill_on_drop: false }` [INFO] [stderr] Cloning into '/workspace/builds/worker-6-tc2/source'... [INFO] [stderr] done. [INFO] started tweaking git repo https://github.com/comfortablynumb/myllm [INFO] finished tweaking git repo https://github.com/comfortablynumb/myllm [INFO] tweaked toml for git repo https://github.com/comfortablynumb/myllm written to /workspace/builds/worker-6-tc2/source/Cargo.toml [INFO] validating manifest of git repo https://github.com/comfortablynumb/myllm on toolchain beta-2026-04-21 [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+beta-2026-04-21" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }` [INFO] crate git repo https://github.com/comfortablynumb/myllm already has a lockfile, it will not be regenerated [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+beta-2026-04-21" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }` [INFO] [stderr] Blocking waiting for file lock on package cache [INFO] [stderr] Updating crates.io index [INFO] [stderr] Blocking waiting for file lock on package cache [INFO] [stderr] Downloading crates ... [INFO] [stderr] Downloaded base64ct v1.8.1 [INFO] [stderr] Downloaded zune-jpeg v0.5.8 [INFO] [stderr] Downloaded auto-future v1.0.0 [INFO] [stderr] Downloaded openvino-finder v0.7.3 [INFO] [stderr] Downloaded reserve-port v2.3.0 [INFO] [stderr] Downloaded rust-multipart-rfc7578_2 v0.8.0 [INFO] [stderr] Downloaded bytesize v2.3.1 [INFO] [stderr] Downloaded metrics-exporter-prometheus v0.16.2 [INFO] [stderr] Downloaded axum-extra v0.10.3 [INFO] [stderr] Downloaded axum-test v17.3.0 [INFO] [stderr] Downloaded metrics-util v0.19.1 [INFO] [stderr] Downloaded openvino-sys v0.7.3 [INFO] [stderr] Downloaded ort-sys v2.0.0-rc.10 [INFO] [stderr] Downloaded portable-atomic v1.12.0 [INFO] [stderr] Downloaded smallvec v2.0.0-alpha.10 [INFO] [stderr] Downloaded ort v2.0.0-rc.10 [INFO] [stderr] Downloaded openvino v0.7.3 [INFO] [stderr] Downloaded cudarc v0.12.1 [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-6-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-6-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:d429b63d4308055ea97f60fb1d3dfca48854a00942f1bd2ad806beaf015945ec" "/opt/rustwide/cargo-home/bin/cargo" "+beta-2026-04-21" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }` [INFO] [stdout] c823c4fc44c3ff711de6750f90d72cefb33f64486b32549af63bc4b283b70ab6 [INFO] running `Command { std: "docker" "start" "-a" "c823c4fc44c3ff711de6750f90d72cefb33f64486b32549af63bc4b283b70ab6", kill_on_drop: false }` [INFO] running `Command { std: "docker" "inspect" "c823c4fc44c3ff711de6750f90d72cefb33f64486b32549af63bc4b283b70ab6", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "c823c4fc44c3ff711de6750f90d72cefb33f64486b32549af63bc4b283b70ab6", kill_on_drop: false }` [INFO] [stdout] c823c4fc44c3ff711de6750f90d72cefb33f64486b32549af63bc4b283b70ab6 [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-6-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-6-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=warn" "-e" "RUSTDOCFLAGS=--cap-lints=warn" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:d429b63d4308055ea97f60fb1d3dfca48854a00942f1bd2ad806beaf015945ec" "/opt/rustwide/cargo-home/bin/cargo" "+beta-2026-04-21" "build" "--frozen" "--message-format=json", kill_on_drop: false }` [INFO] [stdout] ca027f3ac8751d885ad1219b455658330e1e5ef129d030f1fce95bf4ed0c7bb8 [INFO] running `Command { std: "docker" "start" "-a" "ca027f3ac8751d885ad1219b455658330e1e5ef129d030f1fce95bf4ed0c7bb8", kill_on_drop: false }` [INFO] [stderr] Compiling proc-macro2 v1.0.103 [INFO] [stderr] Compiling quote v1.0.42 [INFO] [stderr] Compiling unicode-ident v1.0.22 [INFO] [stderr] Compiling memchr v2.7.6 [INFO] [stderr] Compiling libm v0.2.15 [INFO] [stderr] Compiling zerocopy v0.8.31 [INFO] [stderr] Compiling libc v0.2.178 [INFO] [stderr] Compiling either v1.15.0 [INFO] [stderr] Compiling regex-syntax v0.8.8 [INFO] [stderr] Compiling serde_json v1.0.147 [INFO] [stderr] Compiling portable-atomic v1.12.0 [INFO] [stderr] Compiling unicode-width v0.2.2 [INFO] [stderr] Compiling zmij v0.1.9 [INFO] [stderr] Compiling itoa v1.0.16 [INFO] [stderr] Compiling futures-core v0.3.31 [INFO] [stderr] Compiling futures-sink v0.3.31 [INFO] [stderr] Compiling rayon v1.11.0 [INFO] [stderr] Compiling number_prefix v0.4.0 [INFO] [stderr] Compiling slab v0.4.11 [INFO] [stderr] Compiling futures-io v0.3.31 [INFO] [stderr] Compiling futures-channel v0.3.31 [INFO] [stderr] Compiling hashbrown v0.16.1 [INFO] [stderr] Compiling anstyle-parse v0.2.7 [INFO] [stderr] Compiling rustc-hash v2.1.1 [INFO] [stderr] Compiling colorchoice v1.0.4 [INFO] [stderr] Compiling toml_write v0.1.2 [INFO] [stderr] Compiling anstyle v1.0.13 [INFO] [stderr] Compiling aho-corasick v1.1.4 [INFO] [stderr] Compiling anyhow v1.0.100 [INFO] [stderr] Compiling clap_lex v0.7.6 [INFO] [stderr] Compiling tracing-log v0.2.0 [INFO] [stderr] Compiling anstream v0.6.21 [INFO] [stderr] Compiling getrandom v0.2.16 [INFO] [stderr] Compiling parking_lot_core v0.9.12 [INFO] [stderr] Compiling console v0.15.11 [INFO] [stderr] Compiling syn v2.0.111 [INFO] [stderr] Compiling errno v0.3.14 [INFO] [stderr] Compiling mio v1.1.1 [INFO] [stderr] Compiling rand_core v0.6.4 [INFO] [stderr] Compiling parking_lot v0.12.5 [INFO] [stderr] Compiling signal-hook-registry v1.4.8 [INFO] [stderr] Compiling socket2 v0.6.1 [INFO] [stderr] Compiling indicatif v0.17.11 [INFO] [stderr] Compiling indexmap v2.12.1 [INFO] [stderr] Compiling clap_builder v4.5.53 [INFO] [stderr] Compiling memmap2 v0.9.9 [INFO] [stderr] Compiling num-traits v0.2.19 [INFO] [stderr] Compiling regex-automata v0.4.13 [INFO] [stderr] Compiling ppv-lite86 v0.2.21 [INFO] [stderr] Compiling rand_chacha v0.3.1 [INFO] [stderr] Compiling num-complex v0.4.6 [INFO] [stderr] Compiling num-integer v0.1.46 [INFO] [stderr] Compiling rand v0.8.5 [INFO] [stderr] Compiling ndarray v0.16.1 [INFO] [stderr] Compiling rand_distr v0.4.3 [INFO] [stderr] Compiling regex v1.12.2 [INFO] [stderr] Compiling matchers v0.2.0 [INFO] [stderr] Compiling serde_derive v1.0.228 [INFO] [stderr] Compiling thiserror-impl v1.0.69 [INFO] [stderr] Compiling tracing-attributes v0.1.31 [INFO] [stderr] Compiling futures-macro v0.3.31 [INFO] [stderr] Compiling tokio-macros v2.6.0 [INFO] [stderr] Compiling clap_derive v4.5.49 [INFO] [stderr] Compiling futures-util v0.3.31 [INFO] [stderr] Compiling tokio v1.48.0 [INFO] [stderr] Compiling tracing v0.1.44 [INFO] [stderr] Compiling tracing-subscriber v0.3.22 [INFO] [stderr] Compiling thiserror v1.0.69 [INFO] [stderr] Compiling clap v4.5.53 [INFO] [stderr] Compiling serde v1.0.228 [INFO] [stderr] Compiling myllm-core v0.1.0 (/opt/rustwide/workdir/crates/myllm-core) [INFO] [stderr] Compiling serde_spanned v0.6.9 [INFO] [stderr] Compiling toml_datetime v0.6.11 [INFO] [stderr] Compiling myllm-tokenizer v0.1.0 (/opt/rustwide/workdir/crates/myllm-tokenizer) [INFO] [stderr] Compiling safetensors v0.4.5 [INFO] [stderr] Compiling futures-executor v0.3.31 [INFO] [stderr] Compiling toml_edit v0.22.27 [INFO] [stderr] Compiling futures v0.3.31 [INFO] [stdout] warning: field `config` is never read [INFO] [stdout] --> crates/myllm-core/src/quantization/calibration.rs:283:5 [INFO] [stdout] | [INFO] [stdout] 280 | pub struct ModelCalibrator { [INFO] [stdout] | --------------- field in this struct [INFO] [stdout] ... [INFO] [stdout] 283 | config: CalibratorConfig, [INFO] [stdout] | ^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `ModelCalibrator` has a derived impl for the trait `Debug`, but this is intentionally ignored during dead code analysis [INFO] [stdout] = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: fields `max_abs` and `errors_by_bits` are never read [INFO] [stdout] --> crates/myllm-core/src/quantization/exl2.rs:177:5 [INFO] [stdout] | [INFO] [stdout] 171 | struct GroupStats { [INFO] [stdout] | ---------- fields in this struct [INFO] [stdout] ... [INFO] [stdout] 177 | max_abs: f32, [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 181 | errors_by_bits: Vec, [INFO] [stdout] | ^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `GroupStats` has derived impls for the traits `Clone` and `Debug`, but these are intentionally ignored during dead code analysis [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] Compiling myllm-nn v0.1.0 (/opt/rustwide/workdir/crates/myllm-nn) [INFO] [stdout] warning: unused variable: `fan_in` [INFO] [stdout] --> crates/myllm-nn/src/conv.rs:99:13 [INFO] [stdout] | [INFO] [stdout] 99 | let fan_in = config.in_channels * kh * kw / config.groups; [INFO] [stdout] | ^^^^^^ help: if this is intentional, prefix it with an underscore: `_fan_in` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_variables)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `batch` [INFO] [stdout] --> crates/myllm-nn/src/conv.rs:155:13 [INFO] [stdout] | [INFO] [stdout] 155 | let batch = shape.dim(0); [INFO] [stdout] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_batch` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `sh` [INFO] [stdout] --> crates/myllm-nn/src/conv.rs:161:14 [INFO] [stdout] | [INFO] [stdout] 161 | let (sh, sw) = self.config.stride; [INFO] [stdout] | ^^ help: if this is intentional, prefix it with an underscore: `_sh` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `sw` [INFO] [stdout] --> crates/myllm-nn/src/conv.rs:161:18 [INFO] [stdout] | [INFO] [stdout] 161 | let (sh, sw) = self.config.stride; [INFO] [stdout] | ^^ help: if this is intentional, prefix it with an underscore: `_sw` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `col_size` [INFO] [stdout] --> crates/myllm-nn/src/conv.rs:176:13 [INFO] [stdout] | [INFO] [stdout] 176 | let col_size = channels_per_group * kh * kw; [INFO] [stdout] | ^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_col_size` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `spatial_size` [INFO] [stdout] --> crates/myllm-nn/src/conv.rs:177:13 [INFO] [stdout] | [INFO] [stdout] 177 | let spatial_size = out_h * out_w; [INFO] [stdout] | ^^^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_spatial_size` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: method `unfold` is never used [INFO] [stdout] --> crates/myllm-nn/src/conv.rs:153:8 [INFO] [stdout] | [INFO] [stdout] 95 | impl Conv2d { [INFO] [stdout] | -------------------------- method in this implementation [INFO] [stdout] ... [INFO] [stdout] 153 | fn unfold(&self, input: &Tensor) -> Result> { [INFO] [stdout] | ^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `direction_base` is never read [INFO] [stdout] --> crates/myllm-nn/src/dora.rs:87:5 [INFO] [stdout] | [INFO] [stdout] 83 | pub struct DoraLinear { [INFO] [stdout] | ---------- field in this struct [INFO] [stdout] ... [INFO] [stdout] 87 | direction_base: Tensor, [INFO] [stdout] | ^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `dropout` is never read [INFO] [stdout] --> crates/myllm-nn/src/lora.rs:123:5 [INFO] [stdout] | [INFO] [stdout] 111 | pub struct LoraLinear { [INFO] [stdout] | ---------- field in this struct [INFO] [stdout] ... [INFO] [stdout] 123 | dropout: f32, [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `config` is never read [INFO] [stdout] --> crates/myllm-nn/src/qlora.rs:354:5 [INFO] [stdout] | [INFO] [stdout] 350 | pub struct QLoraAdapter { [INFO] [stdout] | ------------ field in this struct [INFO] [stdout] ... [INFO] [stdout] 354 | config: QLoraConfig, [INFO] [stdout] | ^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] Compiling myllm-model v0.1.0 (/opt/rustwide/workdir/crates/myllm-model) [INFO] [stdout] error[E0433]: cannot find `CoreError` in `myllm_core` [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:729:25 [INFO] [stdout] | [INFO] [stdout] 729 | myllm_core::CoreError::InvalidShape("No pooler output available".to_string()) [INFO] [stdout] | ^^^^^^^^^ could not find `CoreError` in `myllm_core` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0433]: cannot find `CoreError` in `myllm_core` [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:504:25 [INFO] [stdout] | [INFO] [stdout] 504 | myllm_core::CoreError::InvalidShape("No pooler output".to_string()) [INFO] [stdout] | ^^^^^^^^^ could not find `CoreError` in `myllm_core` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `RMSNorm` [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:10:76 [INFO] [stdout] | [INFO] [stdout] 10 | use myllm_nn::{Dropout, Embedding, Layer, LayerNorm, Linear, LinearConfig, RMSNorm}; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `LayerNorm` [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:10:43 [INFO] [stdout] | [INFO] [stdout] 10 | use myllm_nn::{Dropout, Embedding, Layer, LayerNorm, Linear, LinearConfig, RMSNorm}; [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::rope::RotaryEmbedding` [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:15:5 [INFO] [stdout] | [INFO] [stdout] 15 | use crate::rope::RotaryEmbedding; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `unwrap` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:163:56 [INFO] [stdout] | [INFO] [stdout] 163 | Tensor::randn([out_features, in_features]).unwrap() * scale; [INFO] [stdout] | ^^^^^^ method not found in `Tensor<_>` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `unwrap` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:166:48 [INFO] [stdout] | [INFO] [stdout] 166 | Some(Tensor::zeros([out_features]).unwrap()) [INFO] [stdout] | ^^^^^^ method not found in `Tensor<_>` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try` [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:196:21 [INFO] [stdout] | [INFO] [stdout] 196 | let zeros = Tensor::zeros(tensor.shape().dims())?; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ the `?` operator cannot be applied to type `Tensor<_>` [INFO] [stdout] | [INFO] [stdout] = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor<_>` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try` [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:197:20 [INFO] [stdout] | [INFO] [stdout] 197 | let ones = Tensor::ones(tensor.shape().dims())?; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ the `?` operator cannot be applied to type `Tensor<_>` [INFO] [stdout] | [INFO] [stdout] = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor<_>` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: the method `gt` exists for reference `&Tensor`, but its trait bounds were not satisfied [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:200:31 [INFO] [stdout] | [INFO] [stdout] 200 | let positive = tensor.gt(&zeros)?; [INFO] [stdout] | ^^ method cannot be called on `&Tensor` due to unsatisfied trait bounds [INFO] [stdout] | [INFO] [stdout] ::: crates/myllm-core/src/tensor.rs:12:1 [INFO] [stdout] | [INFO] [stdout] 12 | pub struct Tensor { [INFO] [stdout] | ----------------------------------- doesn't satisfy `Tensor: Iterator` or `Tensor: PartialOrd<_>` [INFO] [stdout] | [INFO] [stdout] = note: the following trait bounds were not satisfied: [INFO] [stdout] `Tensor: PartialOrd<_>` [INFO] [stdout] which is required by `&Tensor: PartialOrd<&_>` [INFO] [stdout] `&Tensor: Iterator` [INFO] [stdout] which is required by `&mut &Tensor: Iterator` [INFO] [stdout] `Tensor: Iterator` [INFO] [stdout] which is required by `&mut Tensor: Iterator` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `div_scalar` found for reference `&Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:206:29 [INFO] [stdout] | [INFO] [stdout] 206 | let scaled = tensor.div_scalar(scale)?; [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] help: there is a method `add_scalar` with a similar name [INFO] [stdout] | [INFO] [stdout] 206 - let scaled = tensor.div_scalar(scale)?; [INFO] [stdout] 206 + let scaled = tensor.add_scalar(scale)?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `abs` found for reference `&Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:216:26 [INFO] [stdout] | [INFO] [stdout] 216 | let abs = tensor.abs()?; [INFO] [stdout] | ^^^ method not found in `&Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `div_scalar` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:232:46 [INFO] [stdout] | [INFO] [stdout] 232 | let normalized = self.weight.div_scalar(scale)?; [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] help: there is a method `add_scalar` with a similar name [INFO] [stdout] | [INFO] [stdout] 232 - let normalized = self.weight.div_scalar(scale)?; [INFO] [stdout] 232 + let normalized = self.weight.add_scalar(scale)?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `abs` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:237:46 [INFO] [stdout] | [INFO] [stdout] 237 | let abs_weight = self.weight.abs()?; [INFO] [stdout] | ^^^ method not found in `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `div_scalar` found for reference `&Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:260:40 [INFO] [stdout] | [INFO] [stdout] 260 | let normalized = input.div_scalar(scale)?; [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] help: there is a method `add_scalar` with a similar name [INFO] [stdout] | [INFO] [stdout] 260 - let normalized = input.div_scalar(scale)?; [INFO] [stdout] 260 + let normalized = input.add_scalar(scale)?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this method takes 0 arguments but 2 arguments were supplied [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:273:53 [INFO] [stdout] | [INFO] [stdout] 273 | let output = input.matmul(&quantized_weight.transpose(0, 1)?)?; [INFO] [stdout] | ^^^^^^^^^ - - unexpected argument #2 of type `{integer}` [INFO] [stdout] | | [INFO] [stdout] | unexpected argument #1 of type `{integer}` [INFO] [stdout] | [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:149:12 [INFO] [stdout] | [INFO] [stdout] 149 | pub fn transpose(&self) -> Result { [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] help: remove the extra arguments [INFO] [stdout] | [INFO] [stdout] 273 - let output = input.matmul(&quantized_weight.transpose(0, 1)?)?; [INFO] [stdout] 273 + let output = input.matmul(&quantized_weight.transpose()?)?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `unwrap` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:329:49 [INFO] [stdout] | [INFO] [stdout] 329 | weight: Tensor::ones([hidden_size]).unwrap(), [INFO] [stdout] | ^^^^^^ method not found in `Tensor<_>` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `pow_scalar` found for reference `&Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:336:26 [INFO] [stdout] | [INFO] [stdout] 336 | let variance = x.pow_scalar(2.0)?.mean_dim(-1, true)?; [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] help: there is a method `add_scalar` with a similar name [INFO] [stdout] | [INFO] [stdout] 336 - let variance = x.pow_scalar(2.0)?.mean_dim(-1, true)?; [INFO] [stdout] 336 + let variance = x.add_scalar(2.0)?.mean_dim(-1, true)?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `mul` found for reference `&Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:339:28 [INFO] [stdout] | [INFO] [stdout] 339 | let normalized = x.mul(&rsqrt)?; [INFO] [stdout] | ^^^ this is an associated function, not a method [INFO] [stdout] | [INFO] [stdout] = note: found the following associated functions; to be used as methods, functions must have a `self` parameter [INFO] [stdout] note: the candidate is defined in the trait `Backend` [INFO] [stdout] --> crates/myllm-core/src/backend/mod.rs:64:5 [INFO] [stdout] | [INFO] [stdout] 64 | fn mul(a: &Self::Storage, b: &Self::Storage) -> Result; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] = help: items from traits can only be used if the trait is in scope [INFO] [stdout] help: use associated function syntax instead [INFO] [stdout] | [INFO] [stdout] 339 - let normalized = x.mul(&rsqrt)?; [INFO] [stdout] 339 + let normalized = Tensor::mul(&rsqrt)?; [INFO] [stdout] | [INFO] [stdout] help: trait `Mul` which provides `mul` is implemented but not in scope; perhaps you want to import it [INFO] [stdout] | [INFO] [stdout] 12 + use std::ops::Mul; [INFO] [stdout] | [INFO] [stdout] help: there is a method `matmul` with a similar name [INFO] [stdout] | [INFO] [stdout] 339 | let normalized = x.matmul(&rsqrt)?; [INFO] [stdout] | +++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `mul_scalar` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:429:25 [INFO] [stdout] | [INFO] [stdout] 429 | scores = scores.mul_scalar(self.scale)?; [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] help: there is a method `add_scalar` with a similar name [INFO] [stdout] | [INFO] [stdout] 429 - scores = scores.mul_scalar(self.scale)?; [INFO] [stdout] 429 + scores = scores.add_scalar(self.scale)?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no associated function or constant named `full` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:432:35 [INFO] [stdout] | [INFO] [stdout] 432 | let neg_inf = Tensor::full(scores.shape().dims(), -1e9)?; [INFO] [stdout] | ^^^^ associated function or constant not found in `Tensor<_>` [INFO] [stdout] | [INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions: [INFO] [stdout] Tensor::::from_storage [INFO] [stdout] Tensor::::from_slice [INFO] [stdout] Tensor::::zeros [INFO] [stdout] Tensor::::ones [INFO] [stdout] and 7 others [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:20:5 [INFO] [stdout] | [INFO] [stdout] 20 | pub fn from_storage(storage: B::Storage) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 29 | pub fn from_slice(data: &[f32], shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 35 | pub fn zeros(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 40 | pub fn ones(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `unsqueeze` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:436:43 [INFO] [stdout] | [INFO] [stdout] 436 | let causal_mask = causal_mask.unsqueeze(0)?.unsqueeze(0)?; [INFO] [stdout] | ^^^^^^^^^ method not found in `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no associated function or constant named `full` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:437:35 [INFO] [stdout] | [INFO] [stdout] 437 | let neg_inf = Tensor::full(scores.shape().dims(), -1e9)?; [INFO] [stdout] | ^^^^ associated function or constant not found in `Tensor<_>` [INFO] [stdout] | [INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions: [INFO] [stdout] Tensor::::from_storage [INFO] [stdout] Tensor::::from_slice [INFO] [stdout] Tensor::::zeros [INFO] [stdout] Tensor::::ones [INFO] [stdout] and 7 others [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:20:5 [INFO] [stdout] | [INFO] [stdout] 20 | pub fn from_storage(storage: B::Storage) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 29 | pub fn from_slice(data: &[f32], shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 35 | pub fn zeros(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 40 | pub fn ones(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this method takes 0 arguments but 1 argument was supplied [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:441:35 [INFO] [stdout] | [INFO] [stdout] 441 | let attn_weights = scores.softmax(-1)?; [INFO] [stdout] | ^^^^^^^ -- unexpected argument of type `{integer}` [INFO] [stdout] | [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:189:12 [INFO] [stdout] | [INFO] [stdout] 189 | pub fn softmax(&self) -> Self { [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] help: remove the extra argument [INFO] [stdout] | [INFO] [stdout] 441 - let attn_weights = scores.softmax(-1)?; [INFO] [stdout] 441 + let attn_weights = scores.softmax()?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try` [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:441:28 [INFO] [stdout] | [INFO] [stdout] 441 | let attn_weights = scores.softmax(-1)?; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^ the `?` operator cannot be applied to type `Tensor` [INFO] [stdout] | [INFO] [stdout] = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try` [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:511:34 [INFO] [stdout] | [INFO] [stdout] 511 | let gate_activated = gate.silu()?; [INFO] [stdout] | ^^^^^^^^^^^^ the `?` operator cannot be applied to type `Tensor` [INFO] [stdout] | [INFO] [stdout] = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try` [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:520:13 [INFO] [stdout] | [INFO] [stdout] 520 | gate.gelu()? [INFO] [stdout] | ^^^^^^^^^^^^ the `?` operator cannot be applied to type `Tensor` [INFO] [stdout] | [INFO] [stdout] = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:570:48 [INFO] [stdout] | [INFO] [stdout] 549 | impl BitNetBlock { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 570 | let attn_output = self.dropout.forward(&attn_output)?; [INFO] [stdout] | ------- ^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `add` found for reference `&Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:571:43 [INFO] [stdout] | [INFO] [stdout] 571 | let hidden_states = hidden_states.add(&attn_output)?; [INFO] [stdout] | ^^^ this is an associated function, not a method [INFO] [stdout] | [INFO] [stdout] = note: found the following associated functions; to be used as methods, functions must have a `self` parameter [INFO] [stdout] note: the candidate is defined in the trait `Backend` [INFO] [stdout] --> crates/myllm-core/src/backend/mod.rs:58:5 [INFO] [stdout] | [INFO] [stdout] 58 | fn add(a: &Self::Storage, b: &Self::Storage) -> Result; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] = help: items from traits can only be used if the trait is in scope [INFO] [stdout] help: use associated function syntax instead [INFO] [stdout] | [INFO] [stdout] 571 - let hidden_states = hidden_states.add(&attn_output)?; [INFO] [stdout] 571 + let hidden_states = Tensor::add(&attn_output)?; [INFO] [stdout] | [INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it [INFO] [stdout] | [INFO] [stdout] 12 + use std::ops::Add; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:575:47 [INFO] [stdout] | [INFO] [stdout] 549 | impl BitNetBlock { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 575 | let ffn_output = self.dropout.forward(&ffn_output)?; [INFO] [stdout] | ------- ^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:602:28 [INFO] [stdout] | [INFO] [stdout] 602 | let embed_tokens = Embedding::new(config.vocab_size, config.hidden_size); [INFO] [stdout] | ^^^^^^^^^^^^^^ ----------------- ------------------ unexpected argument #2 of type `usize` [INFO] [stdout] | | [INFO] [stdout] | expected `EmbeddingConfig`, found `usize` [INFO] [stdout] | [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> crates/myllm-nn/src/embedding.rs:35:12 [INFO] [stdout] | [INFO] [stdout] 35 | pub fn new(config: EmbeddingConfig) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: remove the extra argument [INFO] [stdout] | [INFO] [stdout] 602 - let embed_tokens = Embedding::new(config.vocab_size, config.hidden_size); [INFO] [stdout] 602 + let embed_tokens = Embedding::new(/* EmbeddingConfig */); [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:648:53 [INFO] [stdout] | [INFO] [stdout] 648 | params.extend(prefix_params("embed_tokens", self.embed_tokens.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:655:48 [INFO] [stdout] | [INFO] [stdout] 655 | params.extend(prefix_params("lm_head", self.lm_head.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no associated function or constant named `from_vec` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/bitnet.rs:712:13 [INFO] [stdout] | [INFO] [stdout] 712 | Tensor::from_vec(mask, [seq_len, seq_len]) [INFO] [stdout] | ^^^^^^^^ associated function or constant not found in `Tensor<_>` [INFO] [stdout] | [INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions: [INFO] [stdout] Tensor::::from_storage [INFO] [stdout] Tensor::::from_slice [INFO] [stdout] Tensor::::zeros [INFO] [stdout] Tensor::::ones [INFO] [stdout] and 7 others [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:20:5 [INFO] [stdout] | [INFO] [stdout] 20 | pub fn from_storage(storage: B::Storage) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 29 | pub fn from_slice(data: &[f32], shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 35 | pub fn zeros(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 40 | pub fn ones(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] help: there is an associated function `from` with a similar name [INFO] [stdout] --> /rustc/8489f1adbe1c8f8921cf18b9962cf565237a5489/library/core/src/convert/mod.rs:592:4 [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `mul_scalar` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:289:25 [INFO] [stdout] | [INFO] [stdout] 289 | scores = scores.mul_scalar(self.scale)?; [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] help: there is a method `add_scalar` with a similar name [INFO] [stdout] | [INFO] [stdout] 289 - scores = scores.mul_scalar(self.scale)?; [INFO] [stdout] 289 + scores = scores.add_scalar(self.scale)?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no associated function or constant named `full` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:293:35 [INFO] [stdout] | [INFO] [stdout] 293 | let neg_inf = Tensor::full(scores.shape().dims(), -1e9)?; [INFO] [stdout] | ^^^^ associated function or constant not found in `Tensor<_>` [INFO] [stdout] | [INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions: [INFO] [stdout] Tensor::::from_storage [INFO] [stdout] Tensor::::from_slice [INFO] [stdout] Tensor::::zeros [INFO] [stdout] Tensor::::ones [INFO] [stdout] and 7 others [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:20:5 [INFO] [stdout] | [INFO] [stdout] 20 | pub fn from_storage(storage: B::Storage) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 29 | pub fn from_slice(data: &[f32], shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 35 | pub fn zeros(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 40 | pub fn ones(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this method takes 0 arguments but 1 argument was supplied [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:297:35 [INFO] [stdout] | [INFO] [stdout] 297 | let attn_weights = scores.softmax(-1)?; [INFO] [stdout] | ^^^^^^^ -- unexpected argument of type `{integer}` [INFO] [stdout] | [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:189:12 [INFO] [stdout] | [INFO] [stdout] 189 | pub fn softmax(&self) -> Self { [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] help: remove the extra argument [INFO] [stdout] | [INFO] [stdout] 297 - let attn_weights = scores.softmax(-1)?; [INFO] [stdout] 297 + let attn_weights = scores.softmax()?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try` [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:297:28 [INFO] [stdout] | [INFO] [stdout] 297 | let attn_weights = scores.softmax(-1)?; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^ the `?` operator cannot be applied to type `Tensor` [INFO] [stdout] | [INFO] [stdout] = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:300:47 [INFO] [stdout] | [INFO] [stdout] 232 | impl BidirectionalAttention { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 300 | let attn_output = attn_weights.matmul(&v)?; [INFO] [stdout] | ------ ^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:139:12 [INFO] [stdout] | [INFO] [stdout] 139 | pub fn matmul(&self, other: &Self) -> Result { [INFO] [stdout] | ^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:305:29 [INFO] [stdout] | [INFO] [stdout] 232 | impl BidirectionalAttention { [INFO] [stdout] | - expected this type parameter [INFO] [stdout] ... [INFO] [stdout] 305 | self.o_proj.forward(&attn_output) [INFO] [stdout] | ------- ^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:311:47 [INFO] [stdout] | [INFO] [stdout] 311 | params.extend(prefix_params("q_proj", self.q_proj.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:312:47 [INFO] [stdout] | [INFO] [stdout] 312 | params.extend(prefix_params("k_proj", self.k_proj.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:313:47 [INFO] [stdout] | [INFO] [stdout] 313 | params.extend(prefix_params("v_proj", self.v_proj.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:314:47 [INFO] [stdout] | [INFO] [stdout] 314 | params.extend(prefix_params("o_proj", self.o_proj.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try` [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:351:63 [INFO] [stdout] | [INFO] [stdout] 351 | ActivationType::Gelu | ActivationType::GeluNew => h.gelu()?, [INFO] [stdout] | ^^^^^^^^^ the `?` operator cannot be applied to type `Tensor` [INFO] [stdout] | [INFO] [stdout] = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try` [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:352:37 [INFO] [stdout] | [INFO] [stdout] 352 | ActivationType::Relu => h.relu()?, [INFO] [stdout] | ^^^^^^^^^ the `?` operator cannot be applied to type `Tensor` [INFO] [stdout] | [INFO] [stdout] = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try` [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:353:37 [INFO] [stdout] | [INFO] [stdout] 353 | ActivationType::Silu => h.silu()?, [INFO] [stdout] | ^^^^^^^^^ the `?` operator cannot be applied to type `Tensor` [INFO] [stdout] | [INFO] [stdout] = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try` [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:354:37 [INFO] [stdout] | [INFO] [stdout] 354 | ActivationType::Tanh => h.tanh()?, [INFO] [stdout] | ^^^^^^^^^ the `?` operator cannot be applied to type `Tensor` [INFO] [stdout] | [INFO] [stdout] = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:358:30 [INFO] [stdout] | [INFO] [stdout] 327 | impl EncoderFFN { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 358 | self.dropout.forward(&h) [INFO] [stdout] | ------- ^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:358:9 [INFO] [stdout] | [INFO] [stdout] 327 | impl EncoderFFN { [INFO] [stdout] | - expected this type parameter [INFO] [stdout] ... [INFO] [stdout] 347 | pub fn forward(&self, hidden_states: &Tensor) -> Result> { [INFO] [stdout] | ----------------- expected `Result, myllm_core::Error>` because of return type [INFO] [stdout] ... [INFO] [stdout] 358 | self.dropout.forward(&h) [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^ expected `Result, Error>`, found `Result` [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Result, _>` [INFO] [stdout] found enum `Result, _>` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:364:53 [INFO] [stdout] | [INFO] [stdout] 364 | params.extend(prefix_params("intermediate", self.intermediate.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:365:47 [INFO] [stdout] | [INFO] [stdout] 365 | params.extend(prefix_params("output", self.output.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:384:29 [INFO] [stdout] | [INFO] [stdout] 384 | attention_norm: LayerNorm::new(config.hidden_size, config.layer_norm_eps), [INFO] [stdout] | ^^^^^^^^^^^^^^ ------------------ --------------------- unexpected argument #2 of type `f32` [INFO] [stdout] | | [INFO] [stdout] | expected `LayerNormConfig`, found `usize` [INFO] [stdout] | [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> crates/myllm-nn/src/norm.rs:50:12 [INFO] [stdout] | [INFO] [stdout] 50 | pub fn new(config: LayerNormConfig) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: remove the extra argument [INFO] [stdout] | [INFO] [stdout] 384 - attention_norm: LayerNorm::new(config.hidden_size, config.layer_norm_eps), [INFO] [stdout] 384 + attention_norm: LayerNorm::new(/* LayerNormConfig */), [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:386:23 [INFO] [stdout] | [INFO] [stdout] 386 | ffn_norm: LayerNorm::new(config.hidden_size, config.layer_norm_eps), [INFO] [stdout] | ^^^^^^^^^^^^^^ ------------------ --------------------- unexpected argument #2 of type `f32` [INFO] [stdout] | | [INFO] [stdout] | expected `LayerNormConfig`, found `usize` [INFO] [stdout] | [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> crates/myllm-nn/src/norm.rs:50:12 [INFO] [stdout] | [INFO] [stdout] 50 | pub fn new(config: LayerNormConfig) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: remove the extra argument [INFO] [stdout] | [INFO] [stdout] 386 - ffn_norm: LayerNorm::new(config.hidden_size, config.layer_norm_eps), [INFO] [stdout] 386 + ffn_norm: LayerNorm::new(/* LayerNormConfig */), [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:398:48 [INFO] [stdout] | [INFO] [stdout] 379 | impl EncoderLayer { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 398 | let attn_output = self.dropout.forward(&attn_output)?; [INFO] [stdout] | ------- ^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `add` found for reference `&Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:399:43 [INFO] [stdout] | [INFO] [stdout] 399 | let hidden_states = hidden_states.add(&attn_output)?; [INFO] [stdout] | ^^^ this is an associated function, not a method [INFO] [stdout] | [INFO] [stdout] = note: found the following associated functions; to be used as methods, functions must have a `self` parameter [INFO] [stdout] note: the candidate is defined in the trait `Backend` [INFO] [stdout] --> crates/myllm-core/src/backend/mod.rs:58:5 [INFO] [stdout] | [INFO] [stdout] 58 | fn add(a: &Self::Storage, b: &Self::Storage) -> Result; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] = help: items from traits can only be used if the trait is in scope [INFO] [stdout] help: use associated function syntax instead [INFO] [stdout] | [INFO] [stdout] 399 - let hidden_states = hidden_states.add(&attn_output)?; [INFO] [stdout] 399 + let hidden_states = Tensor::add(&attn_output)?; [INFO] [stdout] | [INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it [INFO] [stdout] | [INFO] [stdout] 9 + use std::ops::Add; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `add` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:403:43 [INFO] [stdout] | [INFO] [stdout] 403 | let hidden_states = hidden_states.add(&ffn_output)?; [INFO] [stdout] | ^^^ this is an associated function, not a method [INFO] [stdout] | [INFO] [stdout] --> /rustc/8489f1adbe1c8f8921cf18b9962cf565237a5489/library/core/src/ops/arith.rs:91:7 [INFO] [stdout] | [INFO] [stdout] = note: the method is available for `Tensor` here [INFO] [stdout] | [INFO] [stdout] = note: found the following associated functions; to be used as methods, functions must have a `self` parameter [INFO] [stdout] note: the candidate is defined in the trait `Backend` [INFO] [stdout] --> crates/myllm-core/src/backend/mod.rs:58:5 [INFO] [stdout] | [INFO] [stdout] 58 | fn add(a: &Self::Storage, b: &Self::Storage) -> Result; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] = help: items from traits can only be used if the trait is in scope [INFO] [stdout] help: use associated function syntax instead [INFO] [stdout] | [INFO] [stdout] 403 - let hidden_states = hidden_states.add(&ffn_output)?; [INFO] [stdout] 403 + let hidden_states = Tensor::add(&ffn_output)?; [INFO] [stdout] | [INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it [INFO] [stdout] | [INFO] [stdout] 9 + use std::ops::Add; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:411:55 [INFO] [stdout] | [INFO] [stdout] 411 | params.extend(prefix_params("attention_norm", self.attention_norm.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:413:49 [INFO] [stdout] | [INFO] [stdout] 413 | params.extend(prefix_params("ffn_norm", self.ffn_norm.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:431:30 [INFO] [stdout] | [INFO] [stdout] 431 | word_embeddings: Embedding::new(config.vocab_size, config.hidden_size), [INFO] [stdout] | ^^^^^^^^^^^^^^ ----------------- ------------------ unexpected argument #2 of type `usize` [INFO] [stdout] | | [INFO] [stdout] | expected `EmbeddingConfig`, found `usize` [INFO] [stdout] | [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> crates/myllm-nn/src/embedding.rs:35:12 [INFO] [stdout] | [INFO] [stdout] 35 | pub fn new(config: EmbeddingConfig) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: remove the extra argument [INFO] [stdout] | [INFO] [stdout] 431 - word_embeddings: Embedding::new(config.vocab_size, config.hidden_size), [INFO] [stdout] 431 + word_embeddings: Embedding::new(/* EmbeddingConfig */), [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:432:34 [INFO] [stdout] | [INFO] [stdout] 432 | position_embeddings: Embedding::new(config.max_position_embeddings, config.hidden_size), [INFO] [stdout] | ^^^^^^^^^^^^^^ ------------------------------ ------------------ unexpected argument #2 of type `usize` [INFO] [stdout] | | [INFO] [stdout] | expected `EmbeddingConfig`, found `usize` [INFO] [stdout] | [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> crates/myllm-nn/src/embedding.rs:35:12 [INFO] [stdout] | [INFO] [stdout] 35 | pub fn new(config: EmbeddingConfig) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: remove the extra argument [INFO] [stdout] | [INFO] [stdout] 432 - position_embeddings: Embedding::new(config.max_position_embeddings, config.hidden_size), [INFO] [stdout] 432 + position_embeddings: Embedding::new(/* EmbeddingConfig */), [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:433:36 [INFO] [stdout] | [INFO] [stdout] 433 | token_type_embeddings: Embedding::new(config.type_vocab_size, config.hidden_size), [INFO] [stdout] | ^^^^^^^^^^^^^^ ---------------------- ------------------ unexpected argument #2 of type `usize` [INFO] [stdout] | | [INFO] [stdout] | expected `EmbeddingConfig`, found `usize` [INFO] [stdout] | [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> crates/myllm-nn/src/embedding.rs:35:12 [INFO] [stdout] | [INFO] [stdout] 35 | pub fn new(config: EmbeddingConfig) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: remove the extra argument [INFO] [stdout] | [INFO] [stdout] 433 - token_type_embeddings: Embedding::new(config.type_vocab_size, config.hidden_size), [INFO] [stdout] 433 + token_type_embeddings: Embedding::new(/* EmbeddingConfig */), [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:434:25 [INFO] [stdout] | [INFO] [stdout] 434 | layer_norm: LayerNorm::new(config.hidden_size, config.layer_norm_eps), [INFO] [stdout] | ^^^^^^^^^^^^^^ ------------------ --------------------- unexpected argument #2 of type `f32` [INFO] [stdout] | | [INFO] [stdout] | expected `LayerNormConfig`, found `usize` [INFO] [stdout] | [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> crates/myllm-nn/src/norm.rs:50:12 [INFO] [stdout] | [INFO] [stdout] 50 | pub fn new(config: LayerNormConfig) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: remove the extra argument [INFO] [stdout] | [INFO] [stdout] 434 - layer_norm: LayerNorm::new(config.hidden_size, config.layer_norm_eps), [INFO] [stdout] 434 + layer_norm: LayerNorm::new(/* LayerNormConfig */), [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no associated function or constant named `from_vec` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:454:21 [INFO] [stdout] | [INFO] [stdout] 454 | Tensor::from_vec(positions, [1, seq_len])? [INFO] [stdout] | ^^^^^^^^ associated function or constant not found in `Tensor<_>` [INFO] [stdout] | [INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions: [INFO] [stdout] Tensor::::from_storage [INFO] [stdout] Tensor::::from_slice [INFO] [stdout] Tensor::::zeros [INFO] [stdout] Tensor::::ones [INFO] [stdout] and 7 others [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:20:5 [INFO] [stdout] | [INFO] [stdout] 20 | pub fn from_storage(storage: B::Storage) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 29 | pub fn from_slice(data: &[f32], shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 35 | pub fn zeros(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 40 | pub fn ones(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] help: there is an associated function `from` with a similar name [INFO] [stdout] --> /rustc/8489f1adbe1c8f8921cf18b9962cf565237a5489/library/core/src/convert/mod.rs:592:4 [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try` [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:461:13 [INFO] [stdout] | [INFO] [stdout] 461 | Tensor::zeros([input_ids.shape().dim(0), seq_len, word_embeds.shape().dim(2)])? [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ the `?` operator cannot be applied to type `Tensor<_>` [INFO] [stdout] | [INFO] [stdout] = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor<_>` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `add` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:464:38 [INFO] [stdout] | [INFO] [stdout] 464 | let embeddings = word_embeds.add(&position_embeds)?.add(&token_type_embeds)?; [INFO] [stdout] | ^^^ this is an associated function, not a method [INFO] [stdout] | [INFO] [stdout] --> /rustc/8489f1adbe1c8f8921cf18b9962cf565237a5489/library/core/src/ops/arith.rs:91:7 [INFO] [stdout] | [INFO] [stdout] = note: the method is available for `Tensor` here [INFO] [stdout] | [INFO] [stdout] = note: found the following associated functions; to be used as methods, functions must have a `self` parameter [INFO] [stdout] note: the candidate is defined in the trait `Backend` [INFO] [stdout] --> crates/myllm-core/src/backend/mod.rs:58:5 [INFO] [stdout] | [INFO] [stdout] 58 | fn add(a: &Self::Storage, b: &Self::Storage) -> Result; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] = help: items from traits can only be used if the trait is in scope [INFO] [stdout] help: use associated function syntax instead [INFO] [stdout] | [INFO] [stdout] 464 - let embeddings = word_embeds.add(&position_embeds)?.add(&token_type_embeds)?; [INFO] [stdout] 464 + let embeddings = Tensor::add(&position_embeds)?.add(&token_type_embeds)?; [INFO] [stdout] | [INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it [INFO] [stdout] | [INFO] [stdout] 9 + use std::ops::Add; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:466:30 [INFO] [stdout] | [INFO] [stdout] 427 | impl BertEmbeddings { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 466 | self.dropout.forward(&embeddings) [INFO] [stdout] | ------- ^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:466:9 [INFO] [stdout] | [INFO] [stdout] 427 | impl BertEmbeddings { [INFO] [stdout] | - expected this type parameter [INFO] [stdout] ... [INFO] [stdout] 445 | ) -> Result> { [INFO] [stdout] | ----------------- expected `Result, myllm_core::Error>` because of return type [INFO] [stdout] ... [INFO] [stdout] 466 | self.dropout.forward(&embeddings) [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `Result, Error>`, found `Result` [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Result, _>` [INFO] [stdout] found enum `Result, _>` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:472:56 [INFO] [stdout] | [INFO] [stdout] 472 | params.extend(prefix_params("word_embeddings", self.word_embeddings.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:473:60 [INFO] [stdout] | [INFO] [stdout] 473 | params.extend(prefix_params("position_embeddings", self.position_embeddings.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:474:62 [INFO] [stdout] | [INFO] [stdout] 474 | params.extend(prefix_params("token_type_embeddings", self.token_type_embeddings.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:475:51 [INFO] [stdout] | [INFO] [stdout] 475 | params.extend(prefix_params("layer_norm", self.layer_norm.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `squeeze` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:551:21 [INFO] [stdout] | [INFO] [stdout] 551 | cls.squeeze(1) [INFO] [stdout] | ^^^^^^^ method not found in `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `squeeze` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:571:70 [INFO] [stdout] | [INFO] [stdout] 571 | let first = output.last_hidden_state.slice(1, 0, 1)?.squeeze(1)?; [INFO] [stdout] | ^^^^^^^ method not found in `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `squeeze` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:573:87 [INFO] [stdout] | [INFO] [stdout] 573 | let last = output.last_hidden_state.slice(1, last_idx, last_idx + 1)?.squeeze(1)?; [INFO] [stdout] | ^^^^^^^ method not found in `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0600]: cannot apply unary operator `-` to type `usize` [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:574:47 [INFO] [stdout] | [INFO] [stdout] 574 | Tensor::cat(&[&first, &last], -1) [INFO] [stdout] | ^^ cannot apply unary operator `-` [INFO] [stdout] | [INFO] [stdout] = note: unsigned values cannot be negated [INFO] [stdout] help: you may have meant the maximum value of `usize` [INFO] [stdout] | [INFO] [stdout] 574 - Tensor::cat(&[&first, &last], -1) [INFO] [stdout] 574 + Tensor::cat(&[&first, &last], usize::MAX) [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `squeeze` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:631:35 [INFO] [stdout] | [INFO] [stdout] 631 | let cls_token = cls_token.squeeze(1)?; [INFO] [stdout] | ^^^^^^^ method not found in `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:633:9 [INFO] [stdout] | [INFO] [stdout] 629 | pub fn forward(&self, hidden_states: &Tensor) -> Result> { [INFO] [stdout] | ----------------- expected `Result, myllm_core::Error>` because of return type [INFO] [stdout] ... [INFO] [stdout] 633 | pooled.tanh() [INFO] [stdout] | ^^^^^^^^^^^^^ expected `Result, Error>`, found `Tensor` [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Result, myllm_core::Error>` [INFO] [stdout] found struct `Tensor<_>` [INFO] [stdout] help: try wrapping the expression in `Ok` [INFO] [stdout] | [INFO] [stdout] 633 | Ok(pooled.tanh()) [INFO] [stdout] | +++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:638:32 [INFO] [stdout] | [INFO] [stdout] 638 | prefix_params("dense", self.dense.parameters()) [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:661:44 [INFO] [stdout] | [INFO] [stdout] 648 | impl BertClassificationHead { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 661 | let dropped = self.dropout.forward(pooled_output)?; [INFO] [stdout] | ------- ^^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:662:33 [INFO] [stdout] | [INFO] [stdout] 648 | impl BertClassificationHead { [INFO] [stdout] | - expected this type parameter [INFO] [stdout] ... [INFO] [stdout] 662 | self.classifier.forward(&dropped) [INFO] [stdout] | ------- ^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:667:37 [INFO] [stdout] | [INFO] [stdout] 667 | prefix_params("classifier", self.classifier.parameters()) [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:690:44 [INFO] [stdout] | [INFO] [stdout] 677 | impl BertTokenClassificationHead { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 690 | let dropped = self.dropout.forward(sequence_output)?; [INFO] [stdout] | ------- ^^^^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:691:33 [INFO] [stdout] | [INFO] [stdout] 677 | impl BertTokenClassificationHead { [INFO] [stdout] | - expected this type parameter [INFO] [stdout] ... [INFO] [stdout] 691 | self.classifier.forward(&dropped) [INFO] [stdout] | ------- ^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:696:37 [INFO] [stdout] | [INFO] [stdout] 696 | prefix_params("classifier", self.classifier.parameters()) [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `unsqueeze` found for reference `&Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:796:40 [INFO] [stdout] | [INFO] [stdout] 796 | let mask_expanded = attention_mask.unsqueeze(-1)?; [INFO] [stdout] | ^^^^^^^^^ method not found in `&Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `mul` found for reference `&Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:798:32 [INFO] [stdout] | [INFO] [stdout] 798 | let masked = hidden_states.mul(&mask_expanded)?; [INFO] [stdout] | ^^^ this is an associated function, not a method [INFO] [stdout] | [INFO] [stdout] = note: found the following associated functions; to be used as methods, functions must have a `self` parameter [INFO] [stdout] note: the candidate is defined in the trait `Backend` [INFO] [stdout] --> crates/myllm-core/src/backend/mod.rs:64:5 [INFO] [stdout] | [INFO] [stdout] 64 | fn mul(a: &Self::Storage, b: &Self::Storage) -> Result; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] = help: items from traits can only be used if the trait is in scope [INFO] [stdout] help: use associated function syntax instead [INFO] [stdout] | [INFO] [stdout] 798 - let masked = hidden_states.mul(&mask_expanded)?; [INFO] [stdout] 798 + let masked = Tensor::mul(&mask_expanded)?; [INFO] [stdout] | [INFO] [stdout] help: trait `Mul` which provides `mul` is implemented but not in scope; perhaps you want to import it [INFO] [stdout] | [INFO] [stdout] 9 + use std::ops::Mul; [INFO] [stdout] | [INFO] [stdout] help: there is a method `matmul` with a similar name [INFO] [stdout] | [INFO] [stdout] 798 | let masked = hidden_states.matmul(&mask_expanded)?; [INFO] [stdout] | +++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:188:38 [INFO] [stdout] | [INFO] [stdout] 188 | relative_attention_bias: Embedding::new( [INFO] [stdout] | ^^^^^^^^^^^^^^ [INFO] [stdout] 189 | config.relative_position_buckets, [INFO] [stdout] | -------------------------------- expected `EmbeddingConfig`, found `usize` [INFO] [stdout] 190 | config.num_heads, [INFO] [stdout] | ---------------- unexpected argument #2 of type `usize` [INFO] [stdout] | [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> crates/myllm-nn/src/embedding.rs:35:12 [INFO] [stdout] | [INFO] [stdout] 35 | pub fn new(config: EmbeddingConfig) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: remove the extra argument [INFO] [stdout] | [INFO] [stdout] 189 - config.relative_position_buckets, [INFO] [stdout] 190 - config.num_heads, [INFO] [stdout] 189 + /* EmbeddingConfig */, [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no associated function or constant named `from_vec` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:256:36 [INFO] [stdout] | [INFO] [stdout] 256 | let position_ids = Tensor::from_vec(position_ids, [query_length, key_length])?; [INFO] [stdout] | ^^^^^^^^ associated function or constant not found in `Tensor<_>` [INFO] [stdout] | [INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions: [INFO] [stdout] Tensor::::from_storage [INFO] [stdout] Tensor::::from_slice [INFO] [stdout] Tensor::::zeros [INFO] [stdout] Tensor::::ones [INFO] [stdout] and 7 others [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:20:5 [INFO] [stdout] | [INFO] [stdout] 20 | pub fn from_storage(storage: B::Storage) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 29 | pub fn from_slice(data: &[f32], shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 35 | pub fn zeros(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 40 | pub fn ones(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] help: there is an associated function `from` with a similar name [INFO] [stdout] --> /rustc/8489f1adbe1c8f8921cf18b9962cf565237a5489/library/core/src/convert/mod.rs:592:4 [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `unsqueeze` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:259:35 [INFO] [stdout] | [INFO] [stdout] 259 | bias.permute(&[2, 0, 1])?.unsqueeze(0) [INFO] [stdout] | ^^^^^^^^^ method not found in `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:264:9 [INFO] [stdout] | [INFO] [stdout] 263 | pub fn parameters(&self) -> HashMap> { [INFO] [stdout] | -------------------------- expected `HashMap>` because of return type [INFO] [stdout] 264 | self.relative_attention_bias.parameters() [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `mul_scalar` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:330:25 [INFO] [stdout] | [INFO] [stdout] 330 | scores = scores.mul_scalar(self.scale)?; [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] help: there is a method `add_scalar` with a similar name [INFO] [stdout] | [INFO] [stdout] 330 - scores = scores.mul_scalar(self.scale)?; [INFO] [stdout] 330 + scores = scores.add_scalar(self.scale)?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `add` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:333:29 [INFO] [stdout] | [INFO] [stdout] 333 | scores = scores.add(bias)?; [INFO] [stdout] | ^^^ this is an associated function, not a method [INFO] [stdout] | [INFO] [stdout] --> /rustc/8489f1adbe1c8f8921cf18b9962cf565237a5489/library/core/src/ops/arith.rs:91:7 [INFO] [stdout] | [INFO] [stdout] = note: the method is available for `Tensor` here [INFO] [stdout] | [INFO] [stdout] = note: found the following associated functions; to be used as methods, functions must have a `self` parameter [INFO] [stdout] note: the candidate is defined in the trait `Backend` [INFO] [stdout] --> crates/myllm-core/src/backend/mod.rs:58:5 [INFO] [stdout] | [INFO] [stdout] 58 | fn add(a: &Self::Storage, b: &Self::Storage) -> Result; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] = help: items from traits can only be used if the trait is in scope [INFO] [stdout] help: use associated function syntax instead [INFO] [stdout] | [INFO] [stdout] 333 - scores = scores.add(bias)?; [INFO] [stdout] 333 + scores = Tensor::add(bias)?; [INFO] [stdout] | [INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it [INFO] [stdout] | [INFO] [stdout] 9 + use std::ops::Add; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no associated function or constant named `full` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:338:35 [INFO] [stdout] | [INFO] [stdout] 338 | let neg_inf = Tensor::full(scores.shape().dims(), -1e9)?; [INFO] [stdout] | ^^^^ associated function or constant not found in `Tensor<_>` [INFO] [stdout] | [INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions: [INFO] [stdout] Tensor::::from_storage [INFO] [stdout] Tensor::::from_slice [INFO] [stdout] Tensor::::zeros [INFO] [stdout] Tensor::::ones [INFO] [stdout] and 7 others [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:20:5 [INFO] [stdout] | [INFO] [stdout] 20 | pub fn from_storage(storage: B::Storage) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 29 | pub fn from_slice(data: &[f32], shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 35 | pub fn zeros(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 40 | pub fn ones(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this method takes 0 arguments but 1 argument was supplied [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:342:35 [INFO] [stdout] | [INFO] [stdout] 342 | let attn_weights = scores.softmax(-1)?; [INFO] [stdout] | ^^^^^^^ -- unexpected argument of type `{integer}` [INFO] [stdout] | [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:189:12 [INFO] [stdout] | [INFO] [stdout] 189 | pub fn softmax(&self) -> Self { [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] help: remove the extra argument [INFO] [stdout] | [INFO] [stdout] 342 - let attn_weights = scores.softmax(-1)?; [INFO] [stdout] 342 + let attn_weights = scores.softmax()?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try` [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:342:28 [INFO] [stdout] | [INFO] [stdout] 342 | let attn_weights = scores.softmax(-1)?; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^ the `?` operator cannot be applied to type `Tensor` [INFO] [stdout] | [INFO] [stdout] = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:345:47 [INFO] [stdout] | [INFO] [stdout] 280 | impl EncoderSelfAttention { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 345 | let attn_output = attn_weights.matmul(&v)?; [INFO] [stdout] | ------ ^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:139:12 [INFO] [stdout] | [INFO] [stdout] 139 | pub fn matmul(&self, other: &Self) -> Result { [INFO] [stdout] | ^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:350:29 [INFO] [stdout] | [INFO] [stdout] 280 | impl EncoderSelfAttention { [INFO] [stdout] | - expected this type parameter [INFO] [stdout] ... [INFO] [stdout] 350 | self.o_proj.forward(&attn_output) [INFO] [stdout] | ------- ^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:356:47 [INFO] [stdout] | [INFO] [stdout] 356 | params.extend(prefix_params("q_proj", self.q_proj.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:357:47 [INFO] [stdout] | [INFO] [stdout] 357 | params.extend(prefix_params("k_proj", self.k_proj.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:358:47 [INFO] [stdout] | [INFO] [stdout] 358 | params.extend(prefix_params("v_proj", self.v_proj.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:359:47 [INFO] [stdout] | [INFO] [stdout] 359 | params.extend(prefix_params("o_proj", self.o_proj.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] Compiling toml v0.8.23 [INFO] [stdout] error[E0599]: no method named `mul_scalar` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:438:25 [INFO] [stdout] | [INFO] [stdout] 438 | scores = scores.mul_scalar(self.scale)?; [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] help: there is a method `add_scalar` with a similar name [INFO] [stdout] | [INFO] [stdout] 438 - scores = scores.mul_scalar(self.scale)?; [INFO] [stdout] 438 + scores = scores.add_scalar(self.scale)?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `add` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:441:29 [INFO] [stdout] | [INFO] [stdout] 441 | scores = scores.add(bias)?; [INFO] [stdout] | ^^^ this is an associated function, not a method [INFO] [stdout] | [INFO] [stdout] --> /rustc/8489f1adbe1c8f8921cf18b9962cf565237a5489/library/core/src/ops/arith.rs:91:7 [INFO] [stdout] | [INFO] [stdout] = note: the method is available for `Tensor` here [INFO] [stdout] | [INFO] [stdout] = note: found the following associated functions; to be used as methods, functions must have a `self` parameter [INFO] [stdout] note: the candidate is defined in the trait `Backend` [INFO] [stdout] --> crates/myllm-core/src/backend/mod.rs:58:5 [INFO] [stdout] | [INFO] [stdout] 58 | fn add(a: &Self::Storage, b: &Self::Storage) -> Result; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] = help: items from traits can only be used if the trait is in scope [INFO] [stdout] help: use associated function syntax instead [INFO] [stdout] | [INFO] [stdout] 441 - scores = scores.add(bias)?; [INFO] [stdout] 441 + scores = Tensor::add(bias)?; [INFO] [stdout] | [INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it [INFO] [stdout] | [INFO] [stdout] 9 + use std::ops::Add; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `unsqueeze` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:445:39 [INFO] [stdout] | [INFO] [stdout] 445 | let causal_mask = causal_mask.unsqueeze(0)?.unsqueeze(0)?; [INFO] [stdout] | ^^^^^^^^^ method not found in `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no associated function or constant named `full` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:446:31 [INFO] [stdout] | [INFO] [stdout] 446 | let neg_inf = Tensor::full(scores.shape().dims(), -1e9)?; [INFO] [stdout] | ^^^^ associated function or constant not found in `Tensor<_>` [INFO] [stdout] | [INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions: [INFO] [stdout] Tensor::::from_storage [INFO] [stdout] Tensor::::from_slice [INFO] [stdout] Tensor::::zeros [INFO] [stdout] Tensor::::ones [INFO] [stdout] and 7 others [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:20:5 [INFO] [stdout] | [INFO] [stdout] 20 | pub fn from_storage(storage: B::Storage) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 29 | pub fn from_slice(data: &[f32], shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 35 | pub fn zeros(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 40 | pub fn ones(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this method takes 0 arguments but 1 argument was supplied [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:449:35 [INFO] [stdout] | [INFO] [stdout] 449 | let attn_weights = scores.softmax(-1)?; [INFO] [stdout] | ^^^^^^^ -- unexpected argument of type `{integer}` [INFO] [stdout] | [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:189:12 [INFO] [stdout] | [INFO] [stdout] 189 | pub fn softmax(&self) -> Self { [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] help: remove the extra argument [INFO] [stdout] | [INFO] [stdout] 449 - let attn_weights = scores.softmax(-1)?; [INFO] [stdout] 449 + let attn_weights = scores.softmax()?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try` [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:449:28 [INFO] [stdout] | [INFO] [stdout] 449 | let attn_weights = scores.softmax(-1)?; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^ the `?` operator cannot be applied to type `Tensor` [INFO] [stdout] | [INFO] [stdout] = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:452:47 [INFO] [stdout] | [INFO] [stdout] 376 | impl DecoderSelfAttention { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 452 | let attn_output = attn_weights.matmul(&v)?; [INFO] [stdout] | ------ ^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:139:12 [INFO] [stdout] | [INFO] [stdout] 139 | pub fn matmul(&self, other: &Self) -> Result { [INFO] [stdout] | ^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:457:42 [INFO] [stdout] | [INFO] [stdout] 376 | impl DecoderSelfAttention { [INFO] [stdout] | - expected this type parameter [INFO] [stdout] ... [INFO] [stdout] 457 | let output = self.o_proj.forward(&attn_output)?; [INFO] [stdout] | ------- ^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:464:47 [INFO] [stdout] | [INFO] [stdout] 464 | params.extend(prefix_params("q_proj", self.q_proj.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:465:47 [INFO] [stdout] | [INFO] [stdout] 465 | params.extend(prefix_params("k_proj", self.k_proj.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:466:47 [INFO] [stdout] | [INFO] [stdout] 466 | params.extend(prefix_params("v_proj", self.v_proj.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:467:47 [INFO] [stdout] | [INFO] [stdout] 467 | params.extend(prefix_params("o_proj", self.o_proj.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `mul_scalar` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:551:25 [INFO] [stdout] | [INFO] [stdout] 551 | scores = scores.mul_scalar(self.scale)?; [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] help: there is a method `add_scalar` with a similar name [INFO] [stdout] | [INFO] [stdout] 551 - scores = scores.mul_scalar(self.scale)?; [INFO] [stdout] 551 + scores = scores.add_scalar(self.scale)?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no associated function or constant named `full` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:555:35 [INFO] [stdout] | [INFO] [stdout] 555 | let neg_inf = Tensor::full(scores.shape().dims(), -1e9)?; [INFO] [stdout] | ^^^^ associated function or constant not found in `Tensor<_>` [INFO] [stdout] | [INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions: [INFO] [stdout] Tensor::::from_storage [INFO] [stdout] Tensor::::from_slice [INFO] [stdout] Tensor::::zeros [INFO] [stdout] Tensor::::ones [INFO] [stdout] and 7 others [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:20:5 [INFO] [stdout] | [INFO] [stdout] 20 | pub fn from_storage(storage: B::Storage) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 29 | pub fn from_slice(data: &[f32], shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 35 | pub fn zeros(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 40 | pub fn ones(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this method takes 0 arguments but 1 argument was supplied [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:559:35 [INFO] [stdout] | [INFO] [stdout] 559 | let attn_weights = scores.softmax(-1)?; [INFO] [stdout] | ^^^^^^^ -- unexpected argument of type `{integer}` [INFO] [stdout] | [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:189:12 [INFO] [stdout] | [INFO] [stdout] 189 | pub fn softmax(&self) -> Self { [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] help: remove the extra argument [INFO] [stdout] | [INFO] [stdout] 559 - let attn_weights = scores.softmax(-1)?; [INFO] [stdout] 559 + let attn_weights = scores.softmax()?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try` [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:559:28 [INFO] [stdout] | [INFO] [stdout] 559 | let attn_weights = scores.softmax(-1)?; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^ the `?` operator cannot be applied to type `Tensor` [INFO] [stdout] | [INFO] [stdout] = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:562:47 [INFO] [stdout] | [INFO] [stdout] 484 | impl CrossAttention { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 562 | let attn_output = attn_weights.matmul(&v)?; [INFO] [stdout] | ------ ^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:139:12 [INFO] [stdout] | [INFO] [stdout] 139 | pub fn matmul(&self, other: &Self) -> Result { [INFO] [stdout] | ^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:567:42 [INFO] [stdout] | [INFO] [stdout] 484 | impl CrossAttention { [INFO] [stdout] | - expected this type parameter [INFO] [stdout] ... [INFO] [stdout] 567 | let output = self.o_proj.forward(&attn_output)?; [INFO] [stdout] | ------- ^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:574:47 [INFO] [stdout] | [INFO] [stdout] 574 | params.extend(prefix_params("q_proj", self.q_proj.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:575:47 [INFO] [stdout] | [INFO] [stdout] 575 | params.extend(prefix_params("k_proj", self.k_proj.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:576:47 [INFO] [stdout] | [INFO] [stdout] 576 | params.extend(prefix_params("v_proj", self.v_proj.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:577:47 [INFO] [stdout] | [INFO] [stdout] 577 | params.extend(prefix_params("o_proj", self.o_proj.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try` [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:628:63 [INFO] [stdout] | [INFO] [stdout] 628 | ActivationType::Gelu | ActivationType::GeluNew => h.gelu()?, [INFO] [stdout] | ^^^^^^^^^ the `?` operator cannot be applied to type `Tensor` [INFO] [stdout] | [INFO] [stdout] = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try` [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:629:37 [INFO] [stdout] | [INFO] [stdout] 629 | ActivationType::Relu => h.relu()?, [INFO] [stdout] | ^^^^^^^^^ the `?` operator cannot be applied to type `Tensor` [INFO] [stdout] | [INFO] [stdout] = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try` [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:630:37 [INFO] [stdout] | [INFO] [stdout] 630 | ActivationType::Silu => h.silu()?, [INFO] [stdout] | ^^^^^^^^^ the `?` operator cannot be applied to type `Tensor` [INFO] [stdout] | [INFO] [stdout] = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try` [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:631:37 [INFO] [stdout] | [INFO] [stdout] 631 | ActivationType::Tanh => h.tanh()?, [INFO] [stdout] | ^^^^^^^^^ the `?` operator cannot be applied to type `Tensor` [INFO] [stdout] | [INFO] [stdout] = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `mul` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:637:23 [INFO] [stdout] | [INFO] [stdout] 637 | h = h.mul(&gate)?; [INFO] [stdout] | ^^^ this is an associated function, not a method [INFO] [stdout] | [INFO] [stdout] --> /rustc/8489f1adbe1c8f8921cf18b9962cf565237a5489/library/core/src/ops/arith.rs:337:7 [INFO] [stdout] | [INFO] [stdout] = note: the method is available for `Tensor` here [INFO] [stdout] | [INFO] [stdout] = note: found the following associated functions; to be used as methods, functions must have a `self` parameter [INFO] [stdout] note: the candidate is defined in the trait `Backend` [INFO] [stdout] --> crates/myllm-core/src/backend/mod.rs:64:5 [INFO] [stdout] | [INFO] [stdout] 64 | fn mul(a: &Self::Storage, b: &Self::Storage) -> Result; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] = help: items from traits can only be used if the trait is in scope [INFO] [stdout] help: use associated function syntax instead [INFO] [stdout] | [INFO] [stdout] 637 - h = h.mul(&gate)?; [INFO] [stdout] 637 + h = Tensor::mul(&gate)?; [INFO] [stdout] | [INFO] [stdout] help: trait `Mul` which provides `mul` is implemented but not in scope; perhaps you want to import it [INFO] [stdout] | [INFO] [stdout] 9 + use std::ops::Mul; [INFO] [stdout] | [INFO] [stdout] help: there is a method `matmul` with a similar name [INFO] [stdout] | [INFO] [stdout] 637 | h = h.matmul(&gate)?; [INFO] [stdout] | +++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:641:34 [INFO] [stdout] | [INFO] [stdout] 592 | impl EncoderDecoderFFN { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 641 | h = self.dropout.forward(&h)?; [INFO] [stdout] | ------- ^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: `?` operator has incompatible types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:641:13 [INFO] [stdout] | [INFO] [stdout] 592 | impl EncoderDecoderFFN { [INFO] [stdout] | - expected this type parameter [INFO] [stdout] ... [INFO] [stdout] 641 | h = self.dropout.forward(&h)?; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^ expected `Tensor`, found `Tensor` [INFO] [stdout] | [INFO] [stdout] = note: `?` operator cannot convert from `Tensor` to `Tensor` [INFO] [stdout] = note: expected struct `Tensor` [INFO] [stdout] found struct `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:648:43 [INFO] [stdout] | [INFO] [stdout] 648 | params.extend(prefix_params("wi", self.wi.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:649:43 [INFO] [stdout] | [INFO] [stdout] 649 | params.extend(prefix_params("wo", self.wo.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:652:49 [INFO] [stdout] | [INFO] [stdout] 652 | params.extend(prefix_params("wi_1", wi_1.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:674:34 [INFO] [stdout] | [INFO] [stdout] 674 | self_attention_norm: RMSNorm::new(config.d_model, config.layer_norm_eps), [INFO] [stdout] | ^^^^^^^^^^^^ -------------- --------------------- unexpected argument #2 of type `f32` [INFO] [stdout] | | [INFO] [stdout] | expected `RMSNormConfig`, found `usize` [INFO] [stdout] | [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> crates/myllm-nn/src/norm.rs:185:12 [INFO] [stdout] | [INFO] [stdout] 185 | pub fn new(config: RMSNormConfig) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: remove the extra argument [INFO] [stdout] | [INFO] [stdout] 674 - self_attention_norm: RMSNorm::new(config.d_model, config.layer_norm_eps), [INFO] [stdout] 674 + self_attention_norm: RMSNorm::new(/* RMSNormConfig */), [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:676:23 [INFO] [stdout] | [INFO] [stdout] 676 | ffn_norm: RMSNorm::new(config.d_model, config.layer_norm_eps), [INFO] [stdout] | ^^^^^^^^^^^^ -------------- --------------------- unexpected argument #2 of type `f32` [INFO] [stdout] | | [INFO] [stdout] | expected `RMSNormConfig`, found `usize` [INFO] [stdout] | [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> crates/myllm-nn/src/norm.rs:185:12 [INFO] [stdout] | [INFO] [stdout] 185 | pub fn new(config: RMSNormConfig) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: remove the extra argument [INFO] [stdout] | [INFO] [stdout] 676 - ffn_norm: RMSNorm::new(config.d_model, config.layer_norm_eps), [INFO] [stdout] 676 + ffn_norm: RMSNorm::new(/* RMSNormConfig */), [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:692:52 [INFO] [stdout] | [INFO] [stdout] 669 | impl T5EncoderLayer { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 692 | let attn_output = self.dropout.forward(&attn_output)?; [INFO] [stdout] | ------- ^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `add` found for reference `&Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:693:47 [INFO] [stdout] | [INFO] [stdout] 693 | let hidden_states = hidden_states.add(&attn_output)?; [INFO] [stdout] | ^^^ this is an associated function, not a method [INFO] [stdout] | [INFO] [stdout] = note: found the following associated functions; to be used as methods, functions must have a `self` parameter [INFO] [stdout] note: the candidate is defined in the trait `Backend` [INFO] [stdout] --> crates/myllm-core/src/backend/mod.rs:58:5 [INFO] [stdout] | [INFO] [stdout] 58 | fn add(a: &Self::Storage, b: &Self::Storage) -> Result; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] = help: items from traits can only be used if the trait is in scope [INFO] [stdout] help: use associated function syntax instead [INFO] [stdout] | [INFO] [stdout] 693 - let hidden_states = hidden_states.add(&attn_output)?; [INFO] [stdout] 693 + let hidden_states = Tensor::add(&attn_output)?; [INFO] [stdout] | [INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it [INFO] [stdout] | [INFO] [stdout] 9 + use std::ops::Add; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:697:51 [INFO] [stdout] | [INFO] [stdout] 669 | impl T5EncoderLayer { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 697 | let ffn_output = self.dropout.forward(&ffn_output)?; [INFO] [stdout] | ------- ^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:703:52 [INFO] [stdout] | [INFO] [stdout] 669 | impl T5EncoderLayer { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 703 | let attn_output = self.dropout.forward(&attn_output)?; [INFO] [stdout] | ------- ^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `add` found for reference `&Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:704:47 [INFO] [stdout] | [INFO] [stdout] 704 | let hidden_states = hidden_states.add(&attn_output)?; [INFO] [stdout] | ^^^ this is an associated function, not a method [INFO] [stdout] | [INFO] [stdout] = note: found the following associated functions; to be used as methods, functions must have a `self` parameter [INFO] [stdout] note: the candidate is defined in the trait `Backend` [INFO] [stdout] --> crates/myllm-core/src/backend/mod.rs:58:5 [INFO] [stdout] | [INFO] [stdout] 58 | fn add(a: &Self::Storage, b: &Self::Storage) -> Result; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] = help: items from traits can only be used if the trait is in scope [INFO] [stdout] help: use associated function syntax instead [INFO] [stdout] | [INFO] [stdout] 704 - let hidden_states = hidden_states.add(&attn_output)?; [INFO] [stdout] 704 + let hidden_states = Tensor::add(&attn_output)?; [INFO] [stdout] | [INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it [INFO] [stdout] | [INFO] [stdout] 9 + use std::ops::Add; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:708:51 [INFO] [stdout] | [INFO] [stdout] 669 | impl T5EncoderLayer { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 708 | let ffn_output = self.dropout.forward(&ffn_output)?; [INFO] [stdout] | ------- ^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `add` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:709:47 [INFO] [stdout] | [INFO] [stdout] 709 | let hidden_states = hidden_states.add(&ffn_output)?; [INFO] [stdout] | ^^^ this is an associated function, not a method [INFO] [stdout] | [INFO] [stdout] --> /rustc/8489f1adbe1c8f8921cf18b9962cf565237a5489/library/core/src/ops/arith.rs:91:7 [INFO] [stdout] | [INFO] [stdout] = note: the method is available for `Tensor` here [INFO] [stdout] | [INFO] [stdout] = note: found the following associated functions; to be used as methods, functions must have a `self` parameter [INFO] [stdout] note: the candidate is defined in the trait `Backend` [INFO] [stdout] --> crates/myllm-core/src/backend/mod.rs:58:5 [INFO] [stdout] | [INFO] [stdout] 58 | fn add(a: &Self::Storage, b: &Self::Storage) -> Result; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] = help: items from traits can only be used if the trait is in scope [INFO] [stdout] help: use associated function syntax instead [INFO] [stdout] | [INFO] [stdout] 709 - let hidden_states = hidden_states.add(&ffn_output)?; [INFO] [stdout] 709 + let hidden_states = Tensor::add(&ffn_output)?; [INFO] [stdout] | [INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it [INFO] [stdout] | [INFO] [stdout] 9 + use std::ops::Add; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:720:13 [INFO] [stdout] | [INFO] [stdout] 718 | params.extend(prefix_params( [INFO] [stdout] | ------------- arguments to this function are incorrect [INFO] [stdout] 719 | "self_attention_norm", [INFO] [stdout] 720 | self.self_attention_norm.parameters(), [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:723:49 [INFO] [stdout] | [INFO] [stdout] 723 | params.extend(prefix_params("ffn_norm", self.ffn_norm.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:745:34 [INFO] [stdout] | [INFO] [stdout] 745 | self_attention_norm: RMSNorm::new(config.d_model, config.layer_norm_eps), [INFO] [stdout] | ^^^^^^^^^^^^ -------------- --------------------- unexpected argument #2 of type `f32` [INFO] [stdout] | | [INFO] [stdout] | expected `RMSNormConfig`, found `usize` [INFO] [stdout] | [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> crates/myllm-nn/src/norm.rs:185:12 [INFO] [stdout] | [INFO] [stdout] 185 | pub fn new(config: RMSNormConfig) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: remove the extra argument [INFO] [stdout] | [INFO] [stdout] 745 - self_attention_norm: RMSNorm::new(config.d_model, config.layer_norm_eps), [INFO] [stdout] 745 + self_attention_norm: RMSNorm::new(/* RMSNormConfig */), [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:747:35 [INFO] [stdout] | [INFO] [stdout] 747 | cross_attention_norm: RMSNorm::new(config.d_model, config.layer_norm_eps), [INFO] [stdout] | ^^^^^^^^^^^^ -------------- --------------------- unexpected argument #2 of type `f32` [INFO] [stdout] | | [INFO] [stdout] | expected `RMSNormConfig`, found `usize` [INFO] [stdout] | [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> crates/myllm-nn/src/norm.rs:185:12 [INFO] [stdout] | [INFO] [stdout] 185 | pub fn new(config: RMSNormConfig) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: remove the extra argument [INFO] [stdout] | [INFO] [stdout] 747 - cross_attention_norm: RMSNorm::new(config.d_model, config.layer_norm_eps), [INFO] [stdout] 747 + cross_attention_norm: RMSNorm::new(/* RMSNormConfig */), [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:749:23 [INFO] [stdout] | [INFO] [stdout] 749 | ffn_norm: RMSNorm::new(config.d_model, config.layer_norm_eps), [INFO] [stdout] | ^^^^^^^^^^^^ -------------- --------------------- unexpected argument #2 of type `f32` [INFO] [stdout] | | [INFO] [stdout] | expected `RMSNormConfig`, found `usize` [INFO] [stdout] | [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> crates/myllm-nn/src/norm.rs:185:12 [INFO] [stdout] | [INFO] [stdout] 185 | pub fn new(config: RMSNormConfig) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: remove the extra argument [INFO] [stdout] | [INFO] [stdout] 749 - ffn_norm: RMSNorm::new(config.d_model, config.layer_norm_eps), [INFO] [stdout] 749 + ffn_norm: RMSNorm::new(/* RMSNormConfig */), [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:768:52 [INFO] [stdout] | [INFO] [stdout] 740 | impl T5DecoderLayer { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 768 | let attn_output = self.dropout.forward(&attn_output)?; [INFO] [stdout] | ------- ^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `add` found for reference `&Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:769:35 [INFO] [stdout] | [INFO] [stdout] 769 | let h = hidden_states.add(&attn_output)?; [INFO] [stdout] | ^^^ this is an associated function, not a method [INFO] [stdout] | [INFO] [stdout] = note: found the following associated functions; to be used as methods, functions must have a `self` parameter [INFO] [stdout] note: the candidate is defined in the trait `Backend` [INFO] [stdout] --> crates/myllm-core/src/backend/mod.rs:58:5 [INFO] [stdout] | [INFO] [stdout] 58 | fn add(a: &Self::Storage, b: &Self::Storage) -> Result; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] = help: items from traits can only be used if the trait is in scope [INFO] [stdout] help: use associated function syntax instead [INFO] [stdout] | [INFO] [stdout] 769 - let h = hidden_states.add(&attn_output)?; [INFO] [stdout] 769 + let h = Tensor::add(&attn_output)?; [INFO] [stdout] | [INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it [INFO] [stdout] | [INFO] [stdout] 9 + use std::ops::Add; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:775:52 [INFO] [stdout] | [INFO] [stdout] 740 | impl T5DecoderLayer { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 775 | let attn_output = self.dropout.forward(&attn_output)?; [INFO] [stdout] | ------- ^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `add` found for reference `&Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:776:35 [INFO] [stdout] | [INFO] [stdout] 776 | let h = hidden_states.add(&attn_output)?; [INFO] [stdout] | ^^^ this is an associated function, not a method [INFO] [stdout] | [INFO] [stdout] = note: found the following associated functions; to be used as methods, functions must have a `self` parameter [INFO] [stdout] note: the candidate is defined in the trait `Backend` [INFO] [stdout] --> crates/myllm-core/src/backend/mod.rs:58:5 [INFO] [stdout] | [INFO] [stdout] 58 | fn add(a: &Self::Storage, b: &Self::Storage) -> Result; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] = help: items from traits can only be used if the trait is in scope [INFO] [stdout] help: use associated function syntax instead [INFO] [stdout] | [INFO] [stdout] 776 - let h = hidden_states.add(&attn_output)?; [INFO] [stdout] 776 + let h = Tensor::add(&attn_output)?; [INFO] [stdout] | [INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it [INFO] [stdout] | [INFO] [stdout] 9 + use std::ops::Add; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:789:53 [INFO] [stdout] | [INFO] [stdout] 740 | impl T5DecoderLayer { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 789 | let cross_output = self.dropout.forward(&cross_output)?; [INFO] [stdout] | ------- ^^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `add` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:790:35 [INFO] [stdout] | [INFO] [stdout] 790 | let h = hidden_states.add(&cross_output)?; [INFO] [stdout] | ^^^ this is an associated function, not a method [INFO] [stdout] | [INFO] [stdout] --> /rustc/8489f1adbe1c8f8921cf18b9962cf565237a5489/library/core/src/ops/arith.rs:91:7 [INFO] [stdout] | [INFO] [stdout] = note: the method is available for `Tensor` here [INFO] [stdout] | [INFO] [stdout] = note: found the following associated functions; to be used as methods, functions must have a `self` parameter [INFO] [stdout] note: the candidate is defined in the trait `Backend` [INFO] [stdout] --> crates/myllm-core/src/backend/mod.rs:58:5 [INFO] [stdout] | [INFO] [stdout] 58 | fn add(a: &Self::Storage, b: &Self::Storage) -> Result; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] = help: items from traits can only be used if the trait is in scope [INFO] [stdout] help: use associated function syntax instead [INFO] [stdout] | [INFO] [stdout] 790 - let h = hidden_states.add(&cross_output)?; [INFO] [stdout] 790 + let h = Tensor::add(&cross_output)?; [INFO] [stdout] | [INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it [INFO] [stdout] | [INFO] [stdout] 9 + use std::ops::Add; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:799:53 [INFO] [stdout] | [INFO] [stdout] 740 | impl T5DecoderLayer { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 799 | let cross_output = self.dropout.forward(&cross_output)?; [INFO] [stdout] | ------- ^^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `add` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:800:35 [INFO] [stdout] | [INFO] [stdout] 800 | let h = hidden_states.add(&cross_output)?; [INFO] [stdout] | ^^^ this is an associated function, not a method [INFO] [stdout] | [INFO] [stdout] --> /rustc/8489f1adbe1c8f8921cf18b9962cf565237a5489/library/core/src/ops/arith.rs:91:7 [INFO] [stdout] | [INFO] [stdout] = note: the method is available for `Tensor` here [INFO] [stdout] | [INFO] [stdout] = note: found the following associated functions; to be used as methods, functions must have a `self` parameter [INFO] [stdout] note: the candidate is defined in the trait `Backend` [INFO] [stdout] --> crates/myllm-core/src/backend/mod.rs:58:5 [INFO] [stdout] | [INFO] [stdout] 58 | fn add(a: &Self::Storage, b: &Self::Storage) -> Result; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] = help: items from traits can only be used if the trait is in scope [INFO] [stdout] help: use associated function syntax instead [INFO] [stdout] | [INFO] [stdout] 800 - let h = hidden_states.add(&cross_output)?; [INFO] [stdout] 800 + let h = Tensor::add(&cross_output)?; [INFO] [stdout] | [INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it [INFO] [stdout] | [INFO] [stdout] 9 + use std::ops::Add; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:808:51 [INFO] [stdout] | [INFO] [stdout] 740 | impl T5DecoderLayer { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 808 | let ffn_output = self.dropout.forward(&ffn_output)?; [INFO] [stdout] | ------- ^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `add` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:809:27 [INFO] [stdout] | [INFO] [stdout] 809 | hidden_states.add(&ffn_output)? [INFO] [stdout] | ^^^ this is an associated function, not a method [INFO] [stdout] | [INFO] [stdout] --> /rustc/8489f1adbe1c8f8921cf18b9962cf565237a5489/library/core/src/ops/arith.rs:91:7 [INFO] [stdout] | [INFO] [stdout] = note: the method is available for `Tensor` here [INFO] [stdout] | [INFO] [stdout] = note: found the following associated functions; to be used as methods, functions must have a `self` parameter [INFO] [stdout] note: the candidate is defined in the trait `Backend` [INFO] [stdout] --> crates/myllm-core/src/backend/mod.rs:58:5 [INFO] [stdout] | [INFO] [stdout] 58 | fn add(a: &Self::Storage, b: &Self::Storage) -> Result; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] = help: items from traits can only be used if the trait is in scope [INFO] [stdout] help: use associated function syntax instead [INFO] [stdout] | [INFO] [stdout] 809 - hidden_states.add(&ffn_output)? [INFO] [stdout] 809 + Tensor::add(&ffn_output)? [INFO] [stdout] | [INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it [INFO] [stdout] | [INFO] [stdout] 9 + use std::ops::Add; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:812:51 [INFO] [stdout] | [INFO] [stdout] 740 | impl T5DecoderLayer { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 812 | let ffn_output = self.dropout.forward(&ffn_output)?; [INFO] [stdout] | ------- ^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `add` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:813:35 [INFO] [stdout] | [INFO] [stdout] 813 | let h = hidden_states.add(&ffn_output)?; [INFO] [stdout] | ^^^ this is an associated function, not a method [INFO] [stdout] | [INFO] [stdout] --> /rustc/8489f1adbe1c8f8921cf18b9962cf565237a5489/library/core/src/ops/arith.rs:91:7 [INFO] [stdout] | [INFO] [stdout] = note: the method is available for `Tensor` here [INFO] [stdout] | [INFO] [stdout] = note: found the following associated functions; to be used as methods, functions must have a `self` parameter [INFO] [stdout] note: the candidate is defined in the trait `Backend` [INFO] [stdout] --> crates/myllm-core/src/backend/mod.rs:58:5 [INFO] [stdout] | [INFO] [stdout] 58 | fn add(a: &Self::Storage, b: &Self::Storage) -> Result; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] = help: items from traits can only be used if the trait is in scope [INFO] [stdout] help: use associated function syntax instead [INFO] [stdout] | [INFO] [stdout] 813 - let h = hidden_states.add(&ffn_output)?; [INFO] [stdout] 813 + let h = Tensor::add(&ffn_output)?; [INFO] [stdout] | [INFO] [stdout] help: trait `Add` which provides `add` is implemented but not in scope; perhaps you want to import it [INFO] [stdout] | [INFO] [stdout] 9 + use std::ops::Add; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:831:13 [INFO] [stdout] | [INFO] [stdout] 829 | params.extend(prefix_params( [INFO] [stdout] | ------------- arguments to this function are incorrect [INFO] [stdout] 830 | "self_attention_norm", [INFO] [stdout] 831 | self.self_attention_norm.parameters(), [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:836:13 [INFO] [stdout] | [INFO] [stdout] 834 | params.extend(prefix_params( [INFO] [stdout] | ------------- arguments to this function are incorrect [INFO] [stdout] 835 | "cross_attention_norm", [INFO] [stdout] 836 | self.cross_attention_norm.parameters(), [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:839:49 [INFO] [stdout] | [INFO] [stdout] 839 | params.extend(prefix_params("ffn_norm", self.ffn_norm.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:899:27 [INFO] [stdout] | [INFO] [stdout] 899 | embed_tokens: Embedding::new(config.vocab_size, config.d_model), [INFO] [stdout] | ^^^^^^^^^^^^^^ ----------------- -------------- unexpected argument #2 of type `usize` [INFO] [stdout] | | [INFO] [stdout] | expected `EmbeddingConfig`, found `usize` [INFO] [stdout] | [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> crates/myllm-nn/src/embedding.rs:35:12 [INFO] [stdout] | [INFO] [stdout] 35 | pub fn new(config: EmbeddingConfig) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: remove the extra argument [INFO] [stdout] | [INFO] [stdout] 899 - embed_tokens: Embedding::new(config.vocab_size, config.d_model), [INFO] [stdout] 899 + embed_tokens: Embedding::new(/* EmbeddingConfig */), [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:901:25 [INFO] [stdout] | [INFO] [stdout] 901 | final_norm: RMSNorm::new(config.d_model, config.layer_norm_eps), [INFO] [stdout] | ^^^^^^^^^^^^ -------------- --------------------- unexpected argument #2 of type `f32` [INFO] [stdout] | | [INFO] [stdout] | expected `RMSNormConfig`, found `usize` [INFO] [stdout] | [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> crates/myllm-nn/src/norm.rs:185:12 [INFO] [stdout] | [INFO] [stdout] 185 | pub fn new(config: RMSNormConfig) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: remove the extra argument [INFO] [stdout] | [INFO] [stdout] 901 - final_norm: RMSNorm::new(config.d_model, config.layer_norm_eps), [INFO] [stdout] 901 + final_norm: RMSNorm::new(/* RMSNormConfig */), [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:916:54 [INFO] [stdout] | [INFO] [stdout] 885 | impl T5Encoder { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 916 | let mut hidden_states = self.dropout.forward(&hidden_states)?; [INFO] [stdout] | ------- ^^^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:925:43 [INFO] [stdout] | [INFO] [stdout] 885 | impl T5Encoder { [INFO] [stdout] | - expected this type parameter [INFO] [stdout] ... [INFO] [stdout] 925 | hidden_states = layer.forward(&hidden_states, attention_mask, position_bias.as_ref())?; [INFO] [stdout] | ------- ^^^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:683:12 [INFO] [stdout] | [INFO] [stdout] 683 | pub fn forward( [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] 684 | &self, [INFO] [stdout] 685 | hidden_states: &Tensor, [INFO] [stdout] | ------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: `?` operator has incompatible types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:925:29 [INFO] [stdout] | [INFO] [stdout] 885 | impl T5Encoder { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 925 | hidden_states = layer.forward(&hidden_states, attention_mask, position_bias.as_ref())?; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `Tensor`, found `Tensor` [INFO] [stdout] | [INFO] [stdout] = note: `?` operator cannot convert from `Tensor` to `Tensor` [INFO] [stdout] = note: expected struct `Tensor` [INFO] [stdout] found struct `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:928:33 [INFO] [stdout] | [INFO] [stdout] 885 | impl T5Encoder { [INFO] [stdout] | - expected this type parameter [INFO] [stdout] ... [INFO] [stdout] 928 | self.final_norm.forward(&hidden_states) [INFO] [stdout] | ------- ^^^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:934:53 [INFO] [stdout] | [INFO] [stdout] 934 | params.extend(prefix_params("embed_tokens", self.embed_tokens.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:940:51 [INFO] [stdout] | [INFO] [stdout] 940 | params.extend(prefix_params("final_norm", self.final_norm.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:973:27 [INFO] [stdout] | [INFO] [stdout] 973 | embed_tokens: Embedding::new(config.vocab_size, config.d_model), [INFO] [stdout] | ^^^^^^^^^^^^^^ ----------------- -------------- unexpected argument #2 of type `usize` [INFO] [stdout] | | [INFO] [stdout] | expected `EmbeddingConfig`, found `usize` [INFO] [stdout] | [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> crates/myllm-nn/src/embedding.rs:35:12 [INFO] [stdout] | [INFO] [stdout] 35 | pub fn new(config: EmbeddingConfig) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: remove the extra argument [INFO] [stdout] | [INFO] [stdout] 973 - embed_tokens: Embedding::new(config.vocab_size, config.d_model), [INFO] [stdout] 973 + embed_tokens: Embedding::new(/* EmbeddingConfig */), [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this function takes 1 argument but 2 arguments were supplied [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:975:25 [INFO] [stdout] | [INFO] [stdout] 975 | final_norm: RMSNorm::new(config.d_model, config.layer_norm_eps), [INFO] [stdout] | ^^^^^^^^^^^^ -------------- --------------------- unexpected argument #2 of type `f32` [INFO] [stdout] | | [INFO] [stdout] | expected `RMSNormConfig`, found `usize` [INFO] [stdout] | [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> crates/myllm-nn/src/norm.rs:185:12 [INFO] [stdout] | [INFO] [stdout] 185 | pub fn new(config: RMSNormConfig) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: remove the extra argument [INFO] [stdout] | [INFO] [stdout] 975 - final_norm: RMSNorm::new(config.d_model, config.layer_norm_eps), [INFO] [stdout] 975 + final_norm: RMSNorm::new(/* RMSNormConfig */), [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:997:54 [INFO] [stdout] | [INFO] [stdout] 959 | impl T5Decoder { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 997 | let mut hidden_states = self.dropout.forward(&hidden_states)?; [INFO] [stdout] | ------- ^^^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:1011:17 [INFO] [stdout] | [INFO] [stdout] 959 | impl T5Decoder { [INFO] [stdout] | - expected this type parameter [INFO] [stdout] ... [INFO] [stdout] 1010 | let (output, layer_new_cache) = layer.forward( [INFO] [stdout] | ------- arguments to this method are incorrect [INFO] [stdout] 1011 | &hidden_states, [INFO] [stdout] | ^^^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:756:12 [INFO] [stdout] | [INFO] [stdout] 756 | pub fn forward( [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] 757 | &self, [INFO] [stdout] 758 | hidden_states: &Tensor, [INFO] [stdout] | ------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:1019:29 [INFO] [stdout] | [INFO] [stdout] 959 | impl T5Decoder { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 997 | let mut hidden_states = self.dropout.forward(&hidden_states)?; [INFO] [stdout] | ------------------------------------- expected due to this value [INFO] [stdout] ... [INFO] [stdout] 1019 | hidden_states = output; [INFO] [stdout] | ^^^^^^ expected `Tensor`, found `Tensor` [INFO] [stdout] | [INFO] [stdout] = note: expected struct `Tensor` [INFO] [stdout] found struct `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:1023:53 [INFO] [stdout] | [INFO] [stdout] 959 | impl T5Decoder { [INFO] [stdout] | - expected this type parameter [INFO] [stdout] ... [INFO] [stdout] 1023 | let hidden_states = self.final_norm.forward(&hidden_states)?; [INFO] [stdout] | ------- ^^^^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:1035:53 [INFO] [stdout] | [INFO] [stdout] 1035 | params.extend(prefix_params("embed_tokens", self.embed_tokens.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:1041:51 [INFO] [stdout] | [INFO] [stdout] 1041 | params.extend(prefix_params("final_norm", self.final_norm.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:1138:48 [INFO] [stdout] | [INFO] [stdout] 1138 | params.extend(prefix_params("lm_head", self.lm_head.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no associated function or constant named `full` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:1173:37 [INFO] [stdout] | [INFO] [stdout] 1173 | let decoder_start = Tensor::full([batch_size, 1], self.model.config.decoder_start_token_id as f32)?; [INFO] [stdout] | ^^^^ associated function or constant not found in `Tensor<_>` [INFO] [stdout] | [INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions: [INFO] [stdout] Tensor::::from_storage [INFO] [stdout] Tensor::::from_slice [INFO] [stdout] Tensor::::zeros [INFO] [stdout] Tensor::::ones [INFO] [stdout] and 7 others [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:20:5 [INFO] [stdout] | [INFO] [stdout] 20 | pub fn from_storage(storage: B::Storage) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 29 | pub fn from_slice(data: &[f32], shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 35 | pub fn zeros(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 40 | pub fn ones(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0277]: `DecoderCache` doesn't implement `Debug` [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:1198:5 [INFO] [stdout] | [INFO] [stdout] 1191 | #[derive(Debug)] [INFO] [stdout] | ----- in this derive macro expansion [INFO] [stdout] ... [INFO] [stdout] 1198 | pub cache: DecoderCache, [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^ unsatisfied trait bound [INFO] [stdout] | [INFO] [stdout] help: the trait `Debug` is not implemented for `DecoderCache` [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:859:1 [INFO] [stdout] | [INFO] [stdout] 859 | pub struct DecoderCache { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] = note: add `#[derive(Debug)]` to `DecoderCache` or manually `impl Debug for DecoderCache` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no associated function or constant named `from_vec` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/encoder_decoder.rs:1215:13 [INFO] [stdout] | [INFO] [stdout] 1215 | Tensor::from_vec(mask, [query_len, key_len]) [INFO] [stdout] | ^^^^^^^^ associated function or constant not found in `Tensor<_>` [INFO] [stdout] | [INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions: [INFO] [stdout] Tensor::::from_storage [INFO] [stdout] Tensor::::from_slice [INFO] [stdout] Tensor::::zeros [INFO] [stdout] Tensor::::ones [INFO] [stdout] and 7 others [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:20:5 [INFO] [stdout] | [INFO] [stdout] 20 | pub fn from_storage(storage: B::Storage) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 29 | pub fn from_slice(data: &[f32], shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 35 | pub fn zeros(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 40 | pub fn ones(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] help: there is an associated function `from` with a similar name [INFO] [stdout] --> /rustc/8489f1adbe1c8f8921cf18b9962cf565237a5489/library/core/src/convert/mod.rs:592:4 [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:182:47 [INFO] [stdout] | [INFO] [stdout] 131 | impl DualEncoder { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 182 | let embeddings = self.dropout.forward(&embeddings)?; [INFO] [stdout] | ------- ^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:185:26 [INFO] [stdout] | [INFO] [stdout] 131 | impl DualEncoder { [INFO] [stdout] | - expected this type parameter [INFO] [stdout] ... [INFO] [stdout] 185 | proj.forward(&embeddings)? [INFO] [stdout] | ------- ^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: `if` and `else` have incompatible types [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:187:13 [INFO] [stdout] | [INFO] [stdout] 131 | impl DualEncoder { [INFO] [stdout] | - expected this type parameter [INFO] [stdout] ... [INFO] [stdout] 184 | let embeddings = if let Some(ref proj) = self.query_projection { [INFO] [stdout] | __________________________- [INFO] [stdout] 185 | | proj.forward(&embeddings)? [INFO] [stdout] | | -------------------------- expected because of this [INFO] [stdout] 186 | | } else { [INFO] [stdout] 187 | | embeddings [INFO] [stdout] | | ^^^^^^^^^^ expected `Tensor`, found `Tensor` [INFO] [stdout] 188 | | }; [INFO] [stdout] | |_________- `if` and `else` have incompatible types [INFO] [stdout] | [INFO] [stdout] = note: expected struct `Tensor` [INFO] [stdout] found struct `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:205:47 [INFO] [stdout] | [INFO] [stdout] 131 | impl DualEncoder { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 205 | let embeddings = self.dropout.forward(&embeddings)?; [INFO] [stdout] | ------- ^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:214:26 [INFO] [stdout] | [INFO] [stdout] 131 | impl DualEncoder { [INFO] [stdout] | - expected this type parameter [INFO] [stdout] ... [INFO] [stdout] 214 | proj.forward(&embeddings)? [INFO] [stdout] | ------- ^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: `if` and `else` have incompatible types [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:216:13 [INFO] [stdout] | [INFO] [stdout] 131 | impl DualEncoder { [INFO] [stdout] | - expected this type parameter [INFO] [stdout] ... [INFO] [stdout] 213 | let embeddings = if let Some(proj) = projection { [INFO] [stdout] | __________________________- [INFO] [stdout] 214 | | proj.forward(&embeddings)? [INFO] [stdout] | | -------------------------- expected because of this [INFO] [stdout] 215 | | } else { [INFO] [stdout] 216 | | embeddings [INFO] [stdout] | | ^^^^^^^^^^ expected `Tensor`, found `Tensor` [INFO] [stdout] 217 | | }; [INFO] [stdout] | |_________- `if` and `else` have incompatible types [INFO] [stdout] | [INFO] [stdout] = note: expected struct `Tensor` [INFO] [stdout] found struct `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this method takes 0 arguments but 2 arguments were supplied [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:232:62 [INFO] [stdout] | [INFO] [stdout] 232 | let scores = query_embeddings.matmul(&doc_embeddings.transpose(0, 1)?)?; [INFO] [stdout] | ^^^^^^^^^ - - unexpected argument #2 of type `{integer}` [INFO] [stdout] | | [INFO] [stdout] | unexpected argument #1 of type `{integer}` [INFO] [stdout] | [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:149:12 [INFO] [stdout] | [INFO] [stdout] 149 | pub fn transpose(&self) -> Result { [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] help: remove the extra arguments [INFO] [stdout] | [INFO] [stdout] 232 - let scores = query_embeddings.matmul(&doc_embeddings.transpose(0, 1)?)?; [INFO] [stdout] 232 + let scores = query_embeddings.matmul(&doc_embeddings.transpose()?)?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `div_scalar` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:233:16 [INFO] [stdout] | [INFO] [stdout] 233 | scores.div_scalar(self.config.temperature) [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] help: there is a method `add_scalar` with a similar name [INFO] [stdout] | [INFO] [stdout] 233 - scores.div_scalar(self.config.temperature) [INFO] [stdout] 233 + scores.add_scalar(self.config.temperature) [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:270:61 [INFO] [stdout] | [INFO] [stdout] 270 | params.extend(prefix_params("query_projection", proj.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:274:59 [INFO] [stdout] | [INFO] [stdout] 274 | params.extend(prefix_params("doc_projection", proj.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:357:47 [INFO] [stdout] | [INFO] [stdout] 304 | impl ColBERT { [INFO] [stdout] | - found this type parameter [INFO] [stdout] ... [INFO] [stdout] 357 | let embeddings = self.dropout.forward(&embeddings)?; [INFO] [stdout] | ------- ^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `unsqueeze` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:369:60 [INFO] [stdout] | [INFO] [stdout] 369 | let mask_expanded = mask.slice(1, 0, max_len)?.unsqueeze(-1)?; [INFO] [stdout] | ^^^^^^^^^ method not found in `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `mul` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:370:41 [INFO] [stdout] | [INFO] [stdout] 370 | let embeddings = embeddings.mul(&mask_expanded)?; [INFO] [stdout] | ^^^ this is an associated function, not a method [INFO] [stdout] | [INFO] [stdout] --> /rustc/8489f1adbe1c8f8921cf18b9962cf565237a5489/library/core/src/ops/arith.rs:337:7 [INFO] [stdout] | [INFO] [stdout] = note: the method is available for `Tensor` here [INFO] [stdout] | [INFO] [stdout] = note: found the following associated functions; to be used as methods, functions must have a `self` parameter [INFO] [stdout] note: the candidate is defined in the trait `Backend` [INFO] [stdout] --> crates/myllm-core/src/backend/mod.rs:64:5 [INFO] [stdout] | [INFO] [stdout] 64 | fn mul(a: &Self::Storage, b: &Self::Storage) -> Result; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] = help: items from traits can only be used if the trait is in scope [INFO] [stdout] help: use associated function syntax instead [INFO] [stdout] | [INFO] [stdout] 370 - let embeddings = embeddings.mul(&mask_expanded)?; [INFO] [stdout] 370 + let embeddings = Tensor::mul(&mask_expanded)?; [INFO] [stdout] | [INFO] [stdout] help: trait `Mul` which provides `mul` is implemented but not in scope; perhaps you want to import it [INFO] [stdout] | [INFO] [stdout] 9 + use std::ops::Mul; [INFO] [stdout] | [INFO] [stdout] help: there is a method `matmul` with a similar name [INFO] [stdout] | [INFO] [stdout] 370 | let embeddings = embeddings.matmul(&mask_expanded)?; [INFO] [stdout] | +++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:378:35 [INFO] [stdout] | [INFO] [stdout] 304 | impl ColBERT { [INFO] [stdout] | - expected this type parameter [INFO] [stdout] ... [INFO] [stdout] 378 | normalize_l2_last_dim(&embeddings) [INFO] [stdout] | --------------------- ^^^^^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:785:4 [INFO] [stdout] | [INFO] [stdout] 785 | fn normalize_l2_last_dim(tensor: &Tensor) -> Result> { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^ ------------------ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:380:16 [INFO] [stdout] | [INFO] [stdout] 304 | impl ColBERT { [INFO] [stdout] | - expected this type parameter [INFO] [stdout] ... [INFO] [stdout] 380 | Ok(embeddings) [INFO] [stdout] | -- ^^^^^^^^^^ expected `Tensor`, found `Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this enum variant are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `Tensor` [INFO] [stdout] found struct `Tensor` [INFO] [stdout] help: the type constructed contains `Tensor` due to the type of the argument passed [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:380:13 [INFO] [stdout] | [INFO] [stdout] 380 | Ok(embeddings) [INFO] [stdout] | ^^^----------^ [INFO] [stdout] | | [INFO] [stdout] | this argument influences the type of `Ok` [INFO] [stdout] note: tuple variant defined here [INFO] [stdout] --> /rustc/8489f1adbe1c8f8921cf18b9962cf565237a5489/library/core/src/result.rs:561:4 [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this method takes 0 arguments but 2 arguments were supplied [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:409:35 [INFO] [stdout] | [INFO] [stdout] 409 | let sim = q.matmul(&d.transpose(0, 1)?)?; [INFO] [stdout] | ^^^^^^^^^ - - unexpected argument #2 of type `{integer}` [INFO] [stdout] | | [INFO] [stdout] | unexpected argument #1 of type `{integer}` [INFO] [stdout] | [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:149:12 [INFO] [stdout] | [INFO] [stdout] 149 | pub fn transpose(&self) -> Result { [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] help: remove the extra arguments [INFO] [stdout] | [INFO] [stdout] 409 - let sim = q.matmul(&d.transpose(0, 1)?)?; [INFO] [stdout] 409 + let sim = q.matmul(&d.transpose()?)?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no associated function or constant named `stack` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:415:17 [INFO] [stdout] | [INFO] [stdout] 415 | Tensor::stack(&scores.iter().collect::>(), 0) [INFO] [stdout] | ^^^^^ associated function or constant not found in `Tensor<_>` [INFO] [stdout] | [INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions: [INFO] [stdout] Tensor::::from_storage [INFO] [stdout] Tensor::::from_slice [INFO] [stdout] Tensor::::zeros [INFO] [stdout] Tensor::::ones [INFO] [stdout] and 7 others [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:20:5 [INFO] [stdout] | [INFO] [stdout] 20 | pub fn from_storage(storage: B::Storage) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 29 | pub fn from_slice(data: &[f32], shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 35 | pub fn zeros(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 40 | pub fn ones(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:446:47 [INFO] [stdout] | [INFO] [stdout] 446 | params.extend(prefix_params("linear", self.linear.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:508:46 [INFO] [stdout] | [INFO] [stdout] 474 | impl CrossEncoder { [INFO] [stdout] | - expected this type parameter [INFO] [stdout] ... [INFO] [stdout] 508 | let scores = self.classifier.forward(&pooled)?; [INFO] [stdout] | ------- ^^^^^^^ expected `&Tensor`, found `&Tensor` [INFO] [stdout] | | [INFO] [stdout] | arguments to this method are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected reference `&Tensor` [INFO] [stdout] found reference `&Tensor` [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-nn/src/layer.rs:9:8 [INFO] [stdout] | [INFO] [stdout] 9 | fn forward(&self, input: &Tensor) -> Result>; [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `squeeze` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:510:16 [INFO] [stdout] | [INFO] [stdout] 510 | scores.squeeze(-1) [INFO] [stdout] | ^^^^^^^ method not found in `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:522:51 [INFO] [stdout] | [INFO] [stdout] 522 | params.extend(prefix_params("classifier", self.classifier.parameters())); [INFO] [stdout] | ------------- ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ expected `HashMap>`, found `Vec<&Tensor>` [INFO] [stdout] | | [INFO] [stdout] | arguments to this function are incorrect [INFO] [stdout] | [INFO] [stdout] = note: expected struct `HashMap>` [INFO] [stdout] found struct `Vec<&Tensor>` [INFO] [stdout] note: function defined here [INFO] [stdout] --> crates/myllm-model/src/encoder.rs:808:8 [INFO] [stdout] | [INFO] [stdout] 808 | pub fn prefix_params( [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] 809 | prefix: &str, [INFO] [stdout] 810 | params: HashMap>, [INFO] [stdout] | ---------------------------------- [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `div_scalar` found for reference `&Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:537:36 [INFO] [stdout] | [INFO] [stdout] 537 | let scaled_scores = scores.div_scalar(temperature)?; [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] help: there is a method `add_scalar` with a similar name [INFO] [stdout] | [INFO] [stdout] 537 - let scaled_scores = scores.div_scalar(temperature)?; [INFO] [stdout] 537 + let scaled_scores = scores.add_scalar(temperature)?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no associated function or constant named `from_vec` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:552:17 [INFO] [stdout] | [INFO] [stdout] 552 | Tensor::from_vec(vec![loss_sum / batch_size as f32], [1]) [INFO] [stdout] | ^^^^^^^^ associated function or constant not found in `Tensor<_>` [INFO] [stdout] | [INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions: [INFO] [stdout] Tensor::::from_storage [INFO] [stdout] Tensor::::from_slice [INFO] [stdout] Tensor::::zeros [INFO] [stdout] Tensor::::ones [INFO] [stdout] and 7 others [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:20:5 [INFO] [stdout] | [INFO] [stdout] 20 | pub fn from_storage(storage: B::Storage) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 29 | pub fn from_slice(data: &[f32], shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 35 | pub fn zeros(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 40 | pub fn ones(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] help: there is an associated function `from` with a similar name [INFO] [stdout] --> /rustc/8489f1adbe1c8f8921cf18b9962cf565237a5489/library/core/src/convert/mod.rs:592:4 [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this method takes 0 arguments but 2 arguments were supplied [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:564:57 [INFO] [stdout] | [INFO] [stdout] 564 | let pos_scores = query_emb.matmul(&positive_emb.transpose(0, 1)?)?; [INFO] [stdout] | ^^^^^^^^^ - - unexpected argument #2 of type `{integer}` [INFO] [stdout] | | [INFO] [stdout] | unexpected argument #1 of type `{integer}` [INFO] [stdout] | [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:149:12 [INFO] [stdout] | [INFO] [stdout] 149 | pub fn transpose(&self) -> Result { [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] help: remove the extra arguments [INFO] [stdout] | [INFO] [stdout] 564 - let pos_scores = query_emb.matmul(&positive_emb.transpose(0, 1)?)?; [INFO] [stdout] 564 + let pos_scores = query_emb.matmul(&positive_emb.transpose()?)?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `div_scalar` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:565:37 [INFO] [stdout] | [INFO] [stdout] 565 | let pos_scores = pos_scores.div_scalar(temperature)?; [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] help: there is a method `add_scalar` with a similar name [INFO] [stdout] | [INFO] [stdout] 565 - let pos_scores = pos_scores.div_scalar(temperature)?; [INFO] [stdout] 565 + let pos_scores = pos_scores.add_scalar(temperature)?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this method takes 0 arguments but 2 arguments were supplied [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:568:56 [INFO] [stdout] | [INFO] [stdout] 568 | let neg_scores = query_emb.matmul(&neg_emb.transpose(0, 1)?)?; [INFO] [stdout] | ^^^^^^^^^ - - unexpected argument #2 of type `{integer}` [INFO] [stdout] | | [INFO] [stdout] | unexpected argument #1 of type `{integer}` [INFO] [stdout] | [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:149:12 [INFO] [stdout] | [INFO] [stdout] 149 | pub fn transpose(&self) -> Result { [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] help: remove the extra arguments [INFO] [stdout] | [INFO] [stdout] 568 - let neg_scores = query_emb.matmul(&neg_emb.transpose(0, 1)?)?; [INFO] [stdout] 568 + let neg_scores = query_emb.matmul(&neg_emb.transpose()?)?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `div_scalar` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:569:41 [INFO] [stdout] | [INFO] [stdout] 569 | let neg_scores = neg_scores.div_scalar(temperature)?; [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] help: there is a method `add_scalar` with a similar name [INFO] [stdout] | [INFO] [stdout] 569 - let neg_scores = neg_scores.div_scalar(temperature)?; [INFO] [stdout] 569 + let neg_scores = neg_scores.add_scalar(temperature)?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no associated function or constant named `from_vec` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:590:17 [INFO] [stdout] | [INFO] [stdout] 590 | Tensor::from_vec(vec![loss_sum / batch_size as f32], [1]) [INFO] [stdout] | ^^^^^^^^ associated function or constant not found in `Tensor<_>` [INFO] [stdout] | [INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions: [INFO] [stdout] Tensor::::from_storage [INFO] [stdout] Tensor::::from_slice [INFO] [stdout] Tensor::::zeros [INFO] [stdout] Tensor::::ones [INFO] [stdout] and 7 others [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:20:5 [INFO] [stdout] | [INFO] [stdout] 20 | pub fn from_storage(storage: B::Storage) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 29 | pub fn from_slice(data: &[f32], shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 35 | pub fn zeros(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 40 | pub fn ones(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] help: there is an associated function `from` with a similar name [INFO] [stdout] --> /rustc/8489f1adbe1c8f8921cf18b9962cf565237a5489/library/core/src/convert/mod.rs:592:4 [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `sub` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:603:30 [INFO] [stdout] | [INFO] [stdout] 603 | let loss = (pos_dist.sub(&neg_dist)?.add_scalar(margin)?).relu()?; [INFO] [stdout] | ^^^ this is an associated function, not a method [INFO] [stdout] | [INFO] [stdout] --> /rustc/8489f1adbe1c8f8921cf18b9962cf565237a5489/library/core/src/ops/arith.rs:203:7 [INFO] [stdout] | [INFO] [stdout] = note: the method is available for `Tensor` here [INFO] [stdout] | [INFO] [stdout] = note: found the following associated functions; to be used as methods, functions must have a `self` parameter [INFO] [stdout] note: the candidate is defined in the trait `Backend` [INFO] [stdout] --> crates/myllm-core/src/backend/mod.rs:61:5 [INFO] [stdout] | [INFO] [stdout] 61 | fn sub(a: &Self::Storage, b: &Self::Storage) -> Result; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] = help: items from traits can only be used if the trait is in scope [INFO] [stdout] help: there is a method `sum` with a similar name, but with different arguments [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:219:5 [INFO] [stdout] | [INFO] [stdout] 219 | pub fn sum(&self) -> f32 { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] help: use associated function syntax instead [INFO] [stdout] | [INFO] [stdout] 603 - let loss = (pos_dist.sub(&neg_dist)?.add_scalar(margin)?).relu()?; [INFO] [stdout] 603 + let loss = (Tensor::sub(&neg_dist)?.add_scalar(margin)?).relu()?; [INFO] [stdout] | [INFO] [stdout] help: trait `Sub` which provides `sub` is implemented but not in scope; perhaps you want to import it [INFO] [stdout] | [INFO] [stdout] 9 + use std::ops::Sub; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this method takes 0 arguments but 2 arguments were supplied [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:615:54 [INFO] [stdout] | [INFO] [stdout] 615 | let pos_sim = query_emb.matmul(&positive_emb.transpose(0, 1)?)?; [INFO] [stdout] | ^^^^^^^^^ - - unexpected argument #2 of type `{integer}` [INFO] [stdout] | | [INFO] [stdout] | unexpected argument #1 of type `{integer}` [INFO] [stdout] | [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:149:12 [INFO] [stdout] | [INFO] [stdout] 149 | pub fn transpose(&self) -> Result { [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] help: remove the extra arguments [INFO] [stdout] | [INFO] [stdout] 615 - let pos_sim = query_emb.matmul(&positive_emb.transpose(0, 1)?)?; [INFO] [stdout] 615 + let pos_sim = query_emb.matmul(&positive_emb.transpose()?)?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this method takes 0 arguments but 2 arguments were supplied [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:616:54 [INFO] [stdout] | [INFO] [stdout] 616 | let neg_sim = query_emb.matmul(&negative_emb.transpose(0, 1)?)?; [INFO] [stdout] | ^^^^^^^^^ - - unexpected argument #2 of type `{integer}` [INFO] [stdout] | | [INFO] [stdout] | unexpected argument #1 of type `{integer}` [INFO] [stdout] | [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:149:12 [INFO] [stdout] | [INFO] [stdout] 149 | pub fn transpose(&self) -> Result { [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] help: remove the extra arguments [INFO] [stdout] | [INFO] [stdout] 616 - let neg_sim = query_emb.matmul(&negative_emb.transpose(0, 1)?)?; [INFO] [stdout] 616 + let neg_sim = query_emb.matmul(&negative_emb.transpose()?)?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `sub_scalar` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:621:33 [INFO] [stdout] | [INFO] [stdout] 621 | let delta_pos = pos_sim.sub_scalar(opt_pos)?.neg()?; [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] help: there is a method `add_scalar` with a similar name [INFO] [stdout] | [INFO] [stdout] 621 - let delta_pos = pos_sim.sub_scalar(opt_pos)?.neg()?; [INFO] [stdout] 621 + let delta_pos = pos_sim.add_scalar(opt_pos)?.neg()?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `sub_scalar` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:622:33 [INFO] [stdout] | [INFO] [stdout] 622 | let delta_neg = neg_sim.sub_scalar(opt_neg)?; [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] help: there is a method `add_scalar` with a similar name [INFO] [stdout] | [INFO] [stdout] 622 - let delta_neg = neg_sim.sub_scalar(opt_neg)?; [INFO] [stdout] 622 + let delta_neg = neg_sim.add_scalar(opt_neg)?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `sub_scalar` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:628:27 [INFO] [stdout] | [INFO] [stdout] 628 | .mul(&pos_sim.sub_scalar(1.0 - margin)?)? [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] help: there is a method `add_scalar` with a similar name [INFO] [stdout] | [INFO] [stdout] 628 - .mul(&pos_sim.sub_scalar(1.0 - margin)?)? [INFO] [stdout] 628 + .mul(&pos_sim.add_scalar(1.0 - margin)?)? [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0277]: the `?` operator can only be applied to values that implement `Try` [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:631:19 [INFO] [stdout] | [INFO] [stdout] 631 | .mul(&neg_sim.add_scalar(margin)?)? [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^ the `?` operator cannot be applied to type `Tensor` [INFO] [stdout] | [INFO] [stdout] = help: the nightly-only, unstable trait `Try` is not implemented for `Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `to_vec` found for reference `&Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:677:34 [INFO] [stdout] | [INFO] [stdout] 677 | let scores_data = scores.to_vec()?; [INFO] [stdout] | ^^^^^^ method not found in `&Tensor` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this method takes 0 arguments but 2 arguments were supplied [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:723:62 [INFO] [stdout] | [INFO] [stdout] 723 | let scores = query_embeddings.matmul(&doc_embeddings.transpose(0, 1)?)?; [INFO] [stdout] | ^^^^^^^^^ - - unexpected argument #2 of type `{integer}` [INFO] [stdout] | | [INFO] [stdout] | unexpected argument #1 of type `{integer}` [INFO] [stdout] | [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:149:12 [INFO] [stdout] | [INFO] [stdout] 149 | pub fn transpose(&self) -> Result { [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] help: remove the extra arguments [INFO] [stdout] | [INFO] [stdout] 723 - let scores = query_embeddings.matmul(&doc_embeddings.transpose(0, 1)?)?; [INFO] [stdout] 723 + let scores = query_embeddings.matmul(&doc_embeddings.transpose()?)?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no associated function or constant named `from_vec` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:727:30 [INFO] [stdout] | [INFO] [stdout] 727 | let labels = Tensor::from_vec(labels, [batch_size])?; [INFO] [stdout] | ^^^^^^^^ associated function or constant not found in `Tensor<_>` [INFO] [stdout] | [INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions: [INFO] [stdout] Tensor::::from_storage [INFO] [stdout] Tensor::::from_slice [INFO] [stdout] Tensor::::zeros [INFO] [stdout] Tensor::::ones [INFO] [stdout] and 7 others [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:20:5 [INFO] [stdout] | [INFO] [stdout] 20 | pub fn from_storage(storage: B::Storage) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 29 | pub fn from_slice(data: &[f32], shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 35 | pub fn zeros(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 40 | pub fn ones(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] help: there is an associated function `from` with a similar name [INFO] [stdout] --> /rustc/8489f1adbe1c8f8921cf18b9962cf565237a5489/library/core/src/convert/mod.rs:592:4 [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this method takes 0 arguments but 2 arguments were supplied [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:744:71 [INFO] [stdout] | [INFO] [stdout] 744 | let pos_scores = query_embeddings.matmul(&positive_embeddings.transpose(0, 1)?)?; [INFO] [stdout] | ^^^^^^^^^ - - unexpected argument #2 of type `{integer}` [INFO] [stdout] | | [INFO] [stdout] | unexpected argument #1 of type `{integer}` [INFO] [stdout] | [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:149:12 [INFO] [stdout] | [INFO] [stdout] 149 | pub fn transpose(&self) -> Result { [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] help: remove the extra arguments [INFO] [stdout] | [INFO] [stdout] 744 - let pos_scores = query_embeddings.matmul(&positive_embeddings.transpose(0, 1)?)?; [INFO] [stdout] 744 + let pos_scores = query_embeddings.matmul(&positive_embeddings.transpose()?)?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this method takes 0 arguments but 2 arguments were supplied [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:745:71 [INFO] [stdout] | [INFO] [stdout] 745 | let neg_scores = query_embeddings.matmul(&negative_embeddings.transpose(0, 1)?)?; [INFO] [stdout] | ^^^^^^^^^ - - unexpected argument #2 of type `{integer}` [INFO] [stdout] | | [INFO] [stdout] | unexpected argument #1 of type `{integer}` [INFO] [stdout] | [INFO] [stdout] note: method defined here [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:149:12 [INFO] [stdout] | [INFO] [stdout] 149 | pub fn transpose(&self) -> Result { [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] help: remove the extra arguments [INFO] [stdout] | [INFO] [stdout] 745 - let neg_scores = query_embeddings.matmul(&negative_embeddings.transpose(0, 1)?)?; [INFO] [stdout] 745 + let neg_scores = query_embeddings.matmul(&negative_embeddings.transpose()?)?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no associated function or constant named `from_vec` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:750:30 [INFO] [stdout] | [INFO] [stdout] 750 | let labels = Tensor::from_vec(labels, [batch_size])?; [INFO] [stdout] | ^^^^^^^^ associated function or constant not found in `Tensor<_>` [INFO] [stdout] | [INFO] [stdout] note: if you're trying to build a new `Tensor<_>` consider using one of the following associated functions: [INFO] [stdout] Tensor::::from_storage [INFO] [stdout] Tensor::::from_slice [INFO] [stdout] Tensor::::zeros [INFO] [stdout] Tensor::::ones [INFO] [stdout] and 7 others [INFO] [stdout] --> crates/myllm-core/src/tensor.rs:20:5 [INFO] [stdout] | [INFO] [stdout] 20 | pub fn from_storage(storage: B::Storage) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 29 | pub fn from_slice(data: &[f32], shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 35 | pub fn zeros(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] ... [INFO] [stdout] 40 | pub fn ones(shape: impl Into) -> Self { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] help: there is an associated function `from` with a similar name [INFO] [stdout] --> /rustc/8489f1adbe1c8f8921cf18b9962cf565237a5489/library/core/src/convert/mod.rs:592:4 [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `pow_scalar` found for reference `&Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:779:23 [INFO] [stdout] | [INFO] [stdout] 779 | let norm = tensor.pow_scalar(2.0)?.sum_dim(-1, true)?.sqrt()?; [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] help: there is a method `add_scalar` with a similar name [INFO] [stdout] | [INFO] [stdout] 779 - let norm = tensor.pow_scalar(2.0)?.sum_dim(-1, true)?.sqrt()?; [INFO] [stdout] 779 + let norm = tensor.add_scalar(2.0)?.sum_dim(-1, true)?.sqrt()?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `div` found for reference `&Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:781:12 [INFO] [stdout] | [INFO] [stdout] 781 | tensor.div(&norm) [INFO] [stdout] | ^^^ this is an associated function, not a method [INFO] [stdout] | [INFO] [stdout] = note: found the following associated functions; to be used as methods, functions must have a `self` parameter [INFO] [stdout] note: the candidate is defined in the trait `Backend` [INFO] [stdout] --> crates/myllm-core/src/backend/mod.rs:67:5 [INFO] [stdout] | [INFO] [stdout] 67 | fn div(a: &Self::Storage, b: &Self::Storage) -> Result; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] = help: items from traits can only be used if the trait is in scope [INFO] [stdout] help: use associated function syntax instead [INFO] [stdout] | [INFO] [stdout] 781 - tensor.div(&norm) [INFO] [stdout] 781 + Tensor::div(&norm) [INFO] [stdout] | [INFO] [stdout] help: trait `Div` which provides `div` is implemented but not in scope; perhaps you want to import it [INFO] [stdout] | [INFO] [stdout] 9 + use std::ops::Div; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `pow_scalar` found for reference `&Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:786:23 [INFO] [stdout] | [INFO] [stdout] 786 | let norm = tensor.pow_scalar(2.0)?.sum_dim(-1, true)?.sqrt()?; [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] help: there is a method `add_scalar` with a similar name [INFO] [stdout] | [INFO] [stdout] 786 - let norm = tensor.pow_scalar(2.0)?.sum_dim(-1, true)?.sqrt()?; [INFO] [stdout] 786 + let norm = tensor.add_scalar(2.0)?.sum_dim(-1, true)?.sqrt()?; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `div` found for reference `&Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:788:12 [INFO] [stdout] | [INFO] [stdout] 788 | tensor.div(&norm) [INFO] [stdout] | ^^^ this is an associated function, not a method [INFO] [stdout] | [INFO] [stdout] = note: found the following associated functions; to be used as methods, functions must have a `self` parameter [INFO] [stdout] note: the candidate is defined in the trait `Backend` [INFO] [stdout] --> crates/myllm-core/src/backend/mod.rs:67:5 [INFO] [stdout] | [INFO] [stdout] 67 | fn div(a: &Self::Storage, b: &Self::Storage) -> Result; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] = help: items from traits can only be used if the trait is in scope [INFO] [stdout] help: use associated function syntax instead [INFO] [stdout] | [INFO] [stdout] 788 - tensor.div(&norm) [INFO] [stdout] 788 + Tensor::div(&norm) [INFO] [stdout] | [INFO] [stdout] help: trait `Div` which provides `div` is implemented but not in scope; perhaps you want to import it [INFO] [stdout] | [INFO] [stdout] 9 + use std::ops::Div; [INFO] [stdout] | [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `mul` found for struct `Tensor` in the current scope [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:795:29 [INFO] [stdout] | [INFO] [stdout] 795 | let similarity = a_norm.mul(&b_norm)?.sum_dim(-1, false)?; [INFO] [stdout] | ^^^ this is an associated function, not a method [INFO] [stdout] | [INFO] [stdout] --> /rustc/8489f1adbe1c8f8921cf18b9962cf565237a5489/library/core/src/ops/arith.rs:337:7 [INFO] [stdout] | [INFO] [stdout] = note: the method is available for `Tensor` here [INFO] [stdout] | [INFO] [stdout] = note: found the following associated functions; to be used as methods, functions must have a `self` parameter [INFO] [stdout] note: the candidate is defined in the trait `Backend` [INFO] [stdout] --> crates/myllm-core/src/backend/mod.rs:64:5 [INFO] [stdout] | [INFO] [stdout] 64 | fn mul(a: &Self::Storage, b: &Self::Storage) -> Result; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] = help: items from traits can only be used if the trait is in scope [INFO] [stdout] help: use associated function syntax instead [INFO] [stdout] | [INFO] [stdout] 795 - let similarity = a_norm.mul(&b_norm)?.sum_dim(-1, false)?; [INFO] [stdout] 795 + let similarity = Tensor::mul(&b_norm)?.sum_dim(-1, false)?; [INFO] [stdout] | [INFO] [stdout] help: trait `Mul` which provides `mul` is implemented but not in scope; perhaps you want to import it [INFO] [stdout] | [INFO] [stdout] 9 + use std::ops::Mul; [INFO] [stdout] | [INFO] [stdout] help: there is a method `matmul` with a similar name [INFO] [stdout] | [INFO] [stdout] 795 | let similarity = a_norm.matmul(&b_norm)?.sum_dim(-1, false)?; [INFO] [stdout] | +++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `ratio` [INFO] [stdout] --> crates/myllm-model/src/context_compression.rs:319:13 [INFO] [stdout] | [INFO] [stdout] 319 | let ratio = self.config.compression_ratio; [INFO] [stdout] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_ratio` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_variables)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0382]: use of moved value: `config` [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:171:35 [INFO] [stdout] | [INFO] [stdout] 133 | pub fn new(config: RetrievalConfig) -> Self { [INFO] [stdout] | ------ move occurs because `config` has type `RetrievalConfig`, which does not implement the `Copy` trait [INFO] [stdout] ... [INFO] [stdout] 166 | config, [INFO] [stdout] | ------ value moved here [INFO] [stdout] ... [INFO] [stdout] 171 | dropout: Dropout::new(config.dropout), [INFO] [stdout] | ^^^^^^^^^^^^^^ value used here after move [INFO] [stdout] | [INFO] [stdout] help: consider cloning the value if the performance cost is acceptable [INFO] [stdout] | [INFO] [stdout] 166 | config: config.clone(), [INFO] [stdout] | ++++++++++++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0382]: use of moved value: `config` [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:320:35 [INFO] [stdout] | [INFO] [stdout] 306 | pub fn new(config: RetrievalConfig) -> Self { [INFO] [stdout] | ------ move occurs because `config` has type `RetrievalConfig`, which does not implement the `Copy` trait [INFO] [stdout] ... [INFO] [stdout] 317 | config, [INFO] [stdout] | ------ value moved here [INFO] [stdout] ... [INFO] [stdout] 320 | dropout: Dropout::new(config.dropout), [INFO] [stdout] | ^^^^^^^^^^^^^^ value used here after move [INFO] [stdout] | [INFO] [stdout] help: consider cloning the value if the performance cost is acceptable [INFO] [stdout] | [INFO] [stdout] 317 | config: config.clone(), [INFO] [stdout] | ++++++++++++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0382]: use of moved value: `config` [INFO] [stdout] --> crates/myllm-model/src/retrieval.rs:490:35 [INFO] [stdout] | [INFO] [stdout] 476 | pub fn new(config: RetrievalConfig) -> Self { [INFO] [stdout] | ------ move occurs because `config` has type `RetrievalConfig`, which does not implement the `Copy` trait [INFO] [stdout] ... [INFO] [stdout] 487 | config, [INFO] [stdout] | ------ value moved here [INFO] [stdout] ... [INFO] [stdout] 490 | dropout: Dropout::new(config.dropout), [INFO] [stdout] | ^^^^^^^^^^^^^^ value used here after move [INFO] [stdout] | [INFO] [stdout] help: consider cloning the value if the performance cost is acceptable [INFO] [stdout] | [INFO] [stdout] 487 | config: config.clone(), [INFO] [stdout] | ++++++++++++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] Some errors have detailed explanations: E0061, E0277, E0308, E0382, E0433, E0599, E0600. [INFO] [stdout] [INFO] [stdout] For more information about an error, try `rustc --explain E0061`. [INFO] [stdout] [INFO] [stderr] error: could not compile `myllm-model` (lib) due to 229 previous errors; 4 warnings emitted [INFO] [stderr] warning: build failed, waiting for other jobs to finish... [INFO] running `Command { std: "docker" "inspect" "ca027f3ac8751d885ad1219b455658330e1e5ef129d030f1fce95bf4ed0c7bb8", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "ca027f3ac8751d885ad1219b455658330e1e5ef129d030f1fce95bf4ed0c7bb8", kill_on_drop: false }` [INFO] [stdout] ca027f3ac8751d885ad1219b455658330e1e5ef129d030f1fce95bf4ed0c7bb8