[INFO] cloning repository https://github.com/ano333333/zero-deeplearning [INFO] running `Command { std: "git" "-c" "credential.helper=" "-c" "credential.helper=/workspace/cargo-home/bin/git-credential-null" "clone" "--bare" "https://github.com/ano333333/zero-deeplearning" "/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2Fano333333%2Fzero-deeplearning", kill_on_drop: false }` [INFO] [stderr] Cloning into bare repository '/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2Fano333333%2Fzero-deeplearning'... [INFO] running `Command { std: "git" "rev-parse" "HEAD", kill_on_drop: false }` [INFO] [stdout] 1279e9bd43561db6bb25a6e26fcfac2f3c038558 [INFO] checking ano333333/zero-deeplearning/1279e9bd43561db6bb25a6e26fcfac2f3c038558 against try#4d98531622e1718ae0bc3c7a1ab9cd8938428452+rustflags=-Dtail_expr_drop_order for pr-129604 [INFO] running `Command { std: "git" "clone" "/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2Fano333333%2Fzero-deeplearning" "/workspace/builds/worker-3-tc2/source", kill_on_drop: false }` [INFO] [stderr] Cloning into '/workspace/builds/worker-3-tc2/source'... [INFO] [stderr] done. [INFO] validating manifest of git repo https://github.com/ano333333/zero-deeplearning on toolchain 4d98531622e1718ae0bc3c7a1ab9cd8938428452 [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+4d98531622e1718ae0bc3c7a1ab9cd8938428452" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }` [INFO] started tweaking git repo https://github.com/ano333333/zero-deeplearning [INFO] finished tweaking git repo https://github.com/ano333333/zero-deeplearning [INFO] tweaked toml for git repo https://github.com/ano333333/zero-deeplearning written to /workspace/builds/worker-3-tc2/source/Cargo.toml [INFO] crate git repo https://github.com/ano333333/zero-deeplearning already has a lockfile, it will not be regenerated [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+4d98531622e1718ae0bc3c7a1ab9cd8938428452" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }` [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-3-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-3-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:923055f121b5182466d55868a8b05e67af8ba4a3a3f6bad814e953ca3cd3ac2a" "/opt/rustwide/cargo-home/bin/cargo" "+4d98531622e1718ae0bc3c7a1ab9cd8938428452" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }` [INFO] [stdout] 53289630277af4a32ace340105fc6c977043c571e2273d8358a6ec0626ea39d6 [INFO] running `Command { std: "docker" "start" "-a" "53289630277af4a32ace340105fc6c977043c571e2273d8358a6ec0626ea39d6", kill_on_drop: false }` [INFO] running `Command { std: "docker" "inspect" "53289630277af4a32ace340105fc6c977043c571e2273d8358a6ec0626ea39d6", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "53289630277af4a32ace340105fc6c977043c571e2273d8358a6ec0626ea39d6", kill_on_drop: false }` [INFO] [stdout] 53289630277af4a32ace340105fc6c977043c571e2273d8358a6ec0626ea39d6 [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-3-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-3-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid -Dtail_expr_drop_order" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:923055f121b5182466d55868a8b05e67af8ba4a3a3f6bad814e953ca3cd3ac2a" "/opt/rustwide/cargo-home/bin/cargo" "+4d98531622e1718ae0bc3c7a1ab9cd8938428452" "check" "--frozen" "--all" "--all-targets" "--message-format=json", kill_on_drop: false }` [INFO] [stdout] 8461ee927a8ddf79f6a7aa6d8604a4b38bb3345b115a5126a96a29b736daf891 [INFO] running `Command { std: "docker" "start" "-a" "8461ee927a8ddf79f6a7aa6d8604a4b38bb3345b115a5126a96a29b736daf891", kill_on_drop: false }` [INFO] [stderr] Compiling semver v1.0.23 [INFO] [stderr] Compiling num-traits v0.2.19 [INFO] [stderr] Compiling yeslogic-fontconfig-sys v5.0.0 [INFO] [stderr] Compiling matrixmultiply v0.3.8 [INFO] [stderr] Checking libloading v0.8.3 [INFO] [stderr] Compiling freetype-sys v0.20.1 [INFO] [stderr] Compiling font-kit v0.13.2 [INFO] [stderr] Checking jpeg-decoder v0.3.1 [INFO] [stderr] Checking getrandom v0.2.15 [INFO] [stderr] Checking dirs-sys-next v0.1.2 [INFO] [stderr] Checking plotters-backend v0.3.6 [INFO] [stderr] Checking bytemuck v1.16.0 [INFO] [stderr] Checking weezl v0.1.8 [INFO] [stderr] Compiling cstr v0.2.12 [INFO] [stderr] Checking walkdir v2.5.0 [INFO] [stderr] Checking bitflags v2.5.0 [INFO] [stderr] Checking dirs-next v2.0.0 [INFO] [stderr] Checking dlib v0.5.2 [INFO] [stderr] Checking float-ord v0.3.2 [INFO] [stderr] Checking ttf-parser v0.20.0 [INFO] [stderr] Checking mnist v0.6.0 [INFO] [stderr] Checking gif v0.12.0 [INFO] [stderr] Checking rand_core v0.6.4 [INFO] [stderr] Checking rand_chacha v0.3.1 [INFO] [stderr] Checking plotters-svg v0.3.6 [INFO] [stderr] Compiling rustc_version v0.4.0 [INFO] [stderr] Checking rand v0.8.5 [INFO] [stderr] Checking image v0.24.9 [INFO] [stderr] Checking num-complex v0.4.6 [INFO] [stderr] Checking num-integer v0.1.46 [INFO] [stderr] Checking chrono v0.4.38 [INFO] [stderr] Compiling pathfinder_simd v0.5.3 [INFO] [stderr] Checking ndarray v0.15.6 [INFO] [stderr] Checking rand_distr v0.4.3 [INFO] [stderr] Checking pathfinder_geometry v0.5.1 [INFO] [stderr] Checking plotters-bitmap v0.3.6 [INFO] [stderr] Checking ndarray-rand v0.14.0 [INFO] [stderr] Checking plotters v0.3.6 [INFO] [stderr] Checking zero-deeplearning v0.1.0 (/opt/rustwide/workdir) [INFO] [stdout] warning: unused import: `layer::affine_layer::AffineLayer` [INFO] [stdout] --> src/main.rs:2:5 [INFO] [stdout] | [INFO] [stdout] 2 | use layer::affine_layer::AffineLayer; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `layer::layer::Layer` [INFO] [stdout] --> src/main.rs:3:5 [INFO] [stdout] | [INFO] [stdout] 3 | use layer::layer::Layer; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `layer::relu_layer::ReluLayer` [INFO] [stdout] --> src/main.rs:4:5 [INFO] [stdout] | [INFO] [stdout] 4 | use layer::relu_layer::ReluLayer; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `layer::softmax_with_loss_layer::SoftmaxWithLossLayer` [INFO] [stdout] --> src/main.rs:5:5 [INFO] [stdout] | [INFO] [stdout] 5 | use layer::softmax_with_loss_layer::SoftmaxWithLossLayer; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `two_layer_net::TwoLayerNetGradient` [INFO] [stdout] --> src/main.rs:17:5 [INFO] [stdout] | [INFO] [stdout] 17 | use two_layer_net::TwoLayerNetGradient; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::layer::batch_normalization_layer::BatchNormalizationLayer` [INFO] [stdout] --> src/main.rs:19:5 [INFO] [stdout] | [INFO] [stdout] 19 | use crate::layer::batch_normalization_layer::BatchNormalizationLayer; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `layer::affine_layer::AffineLayer` [INFO] [stdout] --> src/main.rs:2:5 [INFO] [stdout] | [INFO] [stdout] 2 | use layer::affine_layer::AffineLayer; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `layer::layer::Layer` [INFO] [stdout] --> src/main.rs:3:5 [INFO] [stdout] | [INFO] [stdout] 3 | use layer::layer::Layer; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `layer::relu_layer::ReluLayer` [INFO] [stdout] --> src/main.rs:4:5 [INFO] [stdout] | [INFO] [stdout] 4 | use layer::relu_layer::ReluLayer; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `layer::softmax_with_loss_layer::SoftmaxWithLossLayer` [INFO] [stdout] --> src/main.rs:5:5 [INFO] [stdout] | [INFO] [stdout] 5 | use layer::softmax_with_loss_layer::SoftmaxWithLossLayer; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `two_layer_net::TwoLayerNetGradient` [INFO] [stdout] --> src/main.rs:17:5 [INFO] [stdout] | [INFO] [stdout] 17 | use two_layer_net::TwoLayerNetGradient; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `crate::layer::batch_normalization_layer::BatchNormalizationLayer` [INFO] [stdout] --> src/main.rs:19:5 [INFO] [stdout] | [INFO] [stdout] 19 | use crate::layer::batch_normalization_layer::BatchNormalizationLayer; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `du2` [INFO] [stdout] --> src/layer/batch_normalization_layer.rs:59:13 [INFO] [stdout] | [INFO] [stdout] 59 | let du2 = (&dxhat * &self.u1).sum_axis(Axis(1)); [INFO] [stdout] | ^^^ help: if this is intentional, prefix it with an underscore: `_du2` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_variables)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `du2` [INFO] [stdout] --> src/layer/batch_normalization_layer.rs:59:13 [INFO] [stdout] | [INFO] [stdout] 59 | let du2 = (&dxhat * &self.u1).sum_axis(Axis(1)); [INFO] [stdout] | ^^^ help: if this is intentional, prefix it with an underscore: `_du2` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_variables)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: associated function `new` is never used [INFO] [stdout] --> src/layer/add_layer.rs:9:12 [INFO] [stdout] | [INFO] [stdout] 8 | impl AddLayer { [INFO] [stdout] | ---------------------------------- associated function in this implementation [INFO] [stdout] 9 | pub fn new() -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(dead_code)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: associated function `new` is never used [INFO] [stdout] --> src/layer/div_layer.rs:9:12 [INFO] [stdout] | [INFO] [stdout] 8 | impl DivLayer { [INFO] [stdout] | ---------------------------------- associated function in this implementation [INFO] [stdout] 9 | pub fn new() -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: associated function `new` is never used [INFO] [stdout] --> src/layer/exp_layer.rs:9:12 [INFO] [stdout] | [INFO] [stdout] 8 | impl ExpLayer { [INFO] [stdout] | ---------------------------------- associated function in this implementation [INFO] [stdout] 9 | pub fn new() -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: associated function `new` is never used [INFO] [stdout] --> src/layer/mul_layer.rs:10:12 [INFO] [stdout] | [INFO] [stdout] 9 | impl MulLayer { [INFO] [stdout] | ---------------------------------- associated function in this implementation [INFO] [stdout] 10 | pub fn new() -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: associated function `new` is never used [INFO] [stdout] --> src/layer/sigmoid_layer.rs:9:12 [INFO] [stdout] | [INFO] [stdout] 8 | impl SigmoidLayer { [INFO] [stdout] | -------------------------------------- associated function in this implementation [INFO] [stdout] 9 | pub fn new() -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: associated function `new` is never used [INFO] [stdout] --> src/optimize/ada_grad.rs:11:12 [INFO] [stdout] | [INFO] [stdout] 10 | impl AdaGrad { [INFO] [stdout] | ----------------------------- associated function in this implementation [INFO] [stdout] 11 | pub fn new(learning_rate: f64) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: associated function `new` is never used [INFO] [stdout] --> src/optimize/momentum.rs:12:12 [INFO] [stdout] | [INFO] [stdout] 11 | impl Momentum { [INFO] [stdout] | ------------------------------ associated function in this implementation [INFO] [stdout] 12 | pub fn new(learning_rate: f64, momentum: f64) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `identity_function` is never used [INFO] [stdout] --> src/subfunction/identity_function.rs:3:8 [INFO] [stdout] | [INFO] [stdout] 3 | pub fn identity_function(x: ArrayView) -> Array { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `numerical_gradient` is never used [INFO] [stdout] --> src/subfunction/numerical_gradient.rs:3:8 [INFO] [stdout] | [INFO] [stdout] 3 | pub fn numerical_gradient( [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `relu` is never used [INFO] [stdout] --> src/subfunction/relu.rs:3:8 [INFO] [stdout] | [INFO] [stdout] 3 | pub fn relu(x: ArrayView) -> Array { [INFO] [stdout] | ^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `sigmoid` is never used [INFO] [stdout] --> src/subfunction/sigmoid.rs:3:8 [INFO] [stdout] | [INFO] [stdout] 3 | pub fn sigmoid(x: ArrayView) -> Array { [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `softmax` is never used [INFO] [stdout] --> src/subfunction/softmax.rs:3:8 [INFO] [stdout] | [INFO] [stdout] 3 | pub fn softmax(x: ArrayView) -> Array { [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `step_function` is never used [INFO] [stdout] --> src/subfunction/step_function.rs:3:8 [INFO] [stdout] | [INFO] [stdout] 3 | pub fn step_function(x: ArrayView) -> Array { [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error: these values and local bindings have significant drop implementation that will have a different drop order from that of Edition 2021 [INFO] [stdout] --> src/subfunction/cross_entropy_error.rs:7:6 [INFO] [stdout] | [INFO] [stdout] 6 | let log_y = y.mapv(|y| (y + delta).ln()); [INFO] [stdout] | ----- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 7 | -(log_y * t).sum() / batch_size as f64 [INFO] [stdout] | ^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = warning: this changes meaning in Rust 2024 [INFO] [stdout] = note: for more information, see issue #123739 [INFO] [stdout] = note: requested on the command line with `-D tail-expr-drop-order` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error: these values and local bindings have significant drop implementation that will have a different drop order from that of Edition 2021 [INFO] [stdout] --> src/two_layer_net.rs:111:18 [INFO] [stdout] | [INFO] [stdout] 91 | let mut affine1 = self.create_affine1(); [INFO] [stdout] | ------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 92 | let mut batch_normalization1 = self.create_batch_normalization1(); [INFO] [stdout] | -------------------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 93 | let mut relu1 = self.create_relu1(); [INFO] [stdout] | ----- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 94 | let mut affine2 = self.create_affine2(); [INFO] [stdout] | ------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 95 | [INFO] [stdout] 96 | let x = affine1.forward(x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 97 | let x = batch_normalization1.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 98 | let x = relu1.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 99 | let x = affine2.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 100 | [INFO] [stdout] 101 | let mut last_layer = SoftmaxWithLossLayer::new(t); [INFO] [stdout] | ---------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] ... [INFO] [stdout] 105 | let dout = last_layer.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 106 | let dout = affine2.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 107 | let dout = relu1.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] ... [INFO] [stdout] 111 | dw1: affine1.dw.clone(), [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = warning: this changes meaning in Rust 2024 [INFO] [stdout] = note: for more information, see issue #123739 [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error: these values and local bindings have significant drop implementation that will have a different drop order from that of Edition 2021 [INFO] [stdout] --> src/two_layer_net.rs:112:18 [INFO] [stdout] | [INFO] [stdout] 91 | let mut affine1 = self.create_affine1(); [INFO] [stdout] | ------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 92 | let mut batch_normalization1 = self.create_batch_normalization1(); [INFO] [stdout] | -------------------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 93 | let mut relu1 = self.create_relu1(); [INFO] [stdout] | ----- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 94 | let mut affine2 = self.create_affine2(); [INFO] [stdout] | ------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 95 | [INFO] [stdout] 96 | let x = affine1.forward(x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 97 | let x = batch_normalization1.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 98 | let x = relu1.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 99 | let x = affine2.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 100 | [INFO] [stdout] 101 | let mut last_layer = SoftmaxWithLossLayer::new(t); [INFO] [stdout] | ---------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] ... [INFO] [stdout] 105 | let dout = last_layer.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 106 | let dout = affine2.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 107 | let dout = relu1.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] ... [INFO] [stdout] 112 | db1: affine1.db.clone(), [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = warning: this changes meaning in Rust 2024 [INFO] [stdout] = note: for more information, see issue #123739 [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error: these values and local bindings have significant drop implementation that will have a different drop order from that of Edition 2021 [INFO] [stdout] --> src/two_layer_net.rs:113:25 [INFO] [stdout] | [INFO] [stdout] 91 | let mut affine1 = self.create_affine1(); [INFO] [stdout] | ------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 92 | let mut batch_normalization1 = self.create_batch_normalization1(); [INFO] [stdout] | -------------------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 93 | let mut relu1 = self.create_relu1(); [INFO] [stdout] | ----- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 94 | let mut affine2 = self.create_affine2(); [INFO] [stdout] | ------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 95 | [INFO] [stdout] 96 | let x = affine1.forward(x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 97 | let x = batch_normalization1.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 98 | let x = relu1.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 99 | let x = affine2.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 100 | [INFO] [stdout] 101 | let mut last_layer = SoftmaxWithLossLayer::new(t); [INFO] [stdout] | ---------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] ... [INFO] [stdout] 105 | let dout = last_layer.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 106 | let dout = affine2.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 107 | let dout = relu1.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] ... [INFO] [stdout] 113 | dbatch_aff: batch_normalization1.daff.clone(), [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = warning: this changes meaning in Rust 2024 [INFO] [stdout] = note: for more information, see issue #123739 [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error: these values and local bindings have significant drop implementation that will have a different drop order from that of Edition 2021 [INFO] [stdout] --> src/two_layer_net.rs:114:18 [INFO] [stdout] | [INFO] [stdout] 91 | let mut affine1 = self.create_affine1(); [INFO] [stdout] | ------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 92 | let mut batch_normalization1 = self.create_batch_normalization1(); [INFO] [stdout] | -------------------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 93 | let mut relu1 = self.create_relu1(); [INFO] [stdout] | ----- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 94 | let mut affine2 = self.create_affine2(); [INFO] [stdout] | ------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 95 | [INFO] [stdout] 96 | let x = affine1.forward(x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 97 | let x = batch_normalization1.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 98 | let x = relu1.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 99 | let x = affine2.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 100 | [INFO] [stdout] 101 | let mut last_layer = SoftmaxWithLossLayer::new(t); [INFO] [stdout] | ---------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] ... [INFO] [stdout] 105 | let dout = last_layer.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 106 | let dout = affine2.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 107 | let dout = relu1.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] ... [INFO] [stdout] 114 | dw2: affine2.dw.clone(), [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = warning: this changes meaning in Rust 2024 [INFO] [stdout] = note: for more information, see issue #123739 [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error: these values and local bindings have significant drop implementation that will have a different drop order from that of Edition 2021 [INFO] [stdout] --> src/two_layer_net.rs:115:18 [INFO] [stdout] | [INFO] [stdout] 91 | let mut affine1 = self.create_affine1(); [INFO] [stdout] | ------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 92 | let mut batch_normalization1 = self.create_batch_normalization1(); [INFO] [stdout] | -------------------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 93 | let mut relu1 = self.create_relu1(); [INFO] [stdout] | ----- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 94 | let mut affine2 = self.create_affine2(); [INFO] [stdout] | ------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 95 | [INFO] [stdout] 96 | let x = affine1.forward(x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 97 | let x = batch_normalization1.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 98 | let x = relu1.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 99 | let x = affine2.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 100 | [INFO] [stdout] 101 | let mut last_layer = SoftmaxWithLossLayer::new(t); [INFO] [stdout] | ---------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] ... [INFO] [stdout] 105 | let dout = last_layer.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 106 | let dout = affine2.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 107 | let dout = relu1.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] ... [INFO] [stdout] 115 | db2: affine2.db.clone(), [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = warning: this changes meaning in Rust 2024 [INFO] [stdout] = note: for more information, see issue #123739 [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error: aborting due to 6 previous errors; 20 warnings emitted [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] error: could not compile `zero-deeplearning` (bin "zero-deeplearning") due to 7 previous errors; 20 warnings emitted [INFO] [stderr] warning: build failed, waiting for other jobs to finish... [INFO] [stdout] warning: associated function `new` is never used [INFO] [stdout] --> src/layer/add_layer.rs:9:12 [INFO] [stdout] | [INFO] [stdout] 8 | impl AddLayer { [INFO] [stdout] | ---------------------------------- associated function in this implementation [INFO] [stdout] 9 | pub fn new() -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(dead_code)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: associated function `new` is never used [INFO] [stdout] --> src/layer/div_layer.rs:9:12 [INFO] [stdout] | [INFO] [stdout] 8 | impl DivLayer { [INFO] [stdout] | ---------------------------------- associated function in this implementation [INFO] [stdout] 9 | pub fn new() -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: associated function `new` is never used [INFO] [stdout] --> src/layer/exp_layer.rs:9:12 [INFO] [stdout] | [INFO] [stdout] 8 | impl ExpLayer { [INFO] [stdout] | ---------------------------------- associated function in this implementation [INFO] [stdout] 9 | pub fn new() -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: associated function `new` is never used [INFO] [stdout] --> src/layer/mul_layer.rs:10:12 [INFO] [stdout] | [INFO] [stdout] 9 | impl MulLayer { [INFO] [stdout] | ---------------------------------- associated function in this implementation [INFO] [stdout] 10 | pub fn new() -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: associated function `new` is never used [INFO] [stdout] --> src/layer/sigmoid_layer.rs:9:12 [INFO] [stdout] | [INFO] [stdout] 8 | impl SigmoidLayer { [INFO] [stdout] | -------------------------------------- associated function in this implementation [INFO] [stdout] 9 | pub fn new() -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: associated function `new` is never used [INFO] [stdout] --> src/optimize/ada_grad.rs:11:12 [INFO] [stdout] | [INFO] [stdout] 10 | impl AdaGrad { [INFO] [stdout] | ----------------------------- associated function in this implementation [INFO] [stdout] 11 | pub fn new(learning_rate: f64) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: associated function `new` is never used [INFO] [stdout] --> src/optimize/momentum.rs:12:12 [INFO] [stdout] | [INFO] [stdout] 11 | impl Momentum { [INFO] [stdout] | ------------------------------ associated function in this implementation [INFO] [stdout] 12 | pub fn new(learning_rate: f64, momentum: f64) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `identity_function` is never used [INFO] [stdout] --> src/subfunction/identity_function.rs:3:8 [INFO] [stdout] | [INFO] [stdout] 3 | pub fn identity_function(x: ArrayView) -> Array { [INFO] [stdout] | ^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `numerical_gradient` is never used [INFO] [stdout] --> src/subfunction/numerical_gradient.rs:3:8 [INFO] [stdout] | [INFO] [stdout] 3 | pub fn numerical_gradient( [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `relu` is never used [INFO] [stdout] --> src/subfunction/relu.rs:3:8 [INFO] [stdout] | [INFO] [stdout] 3 | pub fn relu(x: ArrayView) -> Array { [INFO] [stdout] | ^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `sigmoid` is never used [INFO] [stdout] --> src/subfunction/sigmoid.rs:3:8 [INFO] [stdout] | [INFO] [stdout] 3 | pub fn sigmoid(x: ArrayView) -> Array { [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `softmax` is never used [INFO] [stdout] --> src/subfunction/softmax.rs:3:8 [INFO] [stdout] | [INFO] [stdout] 3 | pub fn softmax(x: ArrayView) -> Array { [INFO] [stdout] | ^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `step_function` is never used [INFO] [stdout] --> src/subfunction/step_function.rs:3:8 [INFO] [stdout] | [INFO] [stdout] 3 | pub fn step_function(x: ArrayView) -> Array { [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error: these values and local bindings have significant drop implementation that will have a different drop order from that of Edition 2021 [INFO] [stdout] --> src/subfunction/cross_entropy_error.rs:7:6 [INFO] [stdout] | [INFO] [stdout] 6 | let log_y = y.mapv(|y| (y + delta).ln()); [INFO] [stdout] | ----- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 7 | -(log_y * t).sum() / batch_size as f64 [INFO] [stdout] | ^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = warning: this changes meaning in Rust 2024 [INFO] [stdout] = note: for more information, see issue #123739 [INFO] [stdout] = note: requested on the command line with `-D tail-expr-drop-order` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error: these values and local bindings have significant drop implementation that will have a different drop order from that of Edition 2021 [INFO] [stdout] --> src/two_layer_net.rs:111:18 [INFO] [stdout] | [INFO] [stdout] 91 | let mut affine1 = self.create_affine1(); [INFO] [stdout] | ------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 92 | let mut batch_normalization1 = self.create_batch_normalization1(); [INFO] [stdout] | -------------------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 93 | let mut relu1 = self.create_relu1(); [INFO] [stdout] | ----- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 94 | let mut affine2 = self.create_affine2(); [INFO] [stdout] | ------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 95 | [INFO] [stdout] 96 | let x = affine1.forward(x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 97 | let x = batch_normalization1.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 98 | let x = relu1.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 99 | let x = affine2.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 100 | [INFO] [stdout] 101 | let mut last_layer = SoftmaxWithLossLayer::new(t); [INFO] [stdout] | ---------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] ... [INFO] [stdout] 105 | let dout = last_layer.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 106 | let dout = affine2.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 107 | let dout = relu1.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] ... [INFO] [stdout] 111 | dw1: affine1.dw.clone(), [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = warning: this changes meaning in Rust 2024 [INFO] [stdout] = note: for more information, see issue #123739 [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error: these values and local bindings have significant drop implementation that will have a different drop order from that of Edition 2021 [INFO] [stdout] --> src/two_layer_net.rs:112:18 [INFO] [stdout] | [INFO] [stdout] 91 | let mut affine1 = self.create_affine1(); [INFO] [stdout] | ------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 92 | let mut batch_normalization1 = self.create_batch_normalization1(); [INFO] [stdout] | -------------------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 93 | let mut relu1 = self.create_relu1(); [INFO] [stdout] | ----- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 94 | let mut affine2 = self.create_affine2(); [INFO] [stdout] | ------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 95 | [INFO] [stdout] 96 | let x = affine1.forward(x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 97 | let x = batch_normalization1.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 98 | let x = relu1.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 99 | let x = affine2.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 100 | [INFO] [stdout] 101 | let mut last_layer = SoftmaxWithLossLayer::new(t); [INFO] [stdout] | ---------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] ... [INFO] [stdout] 105 | let dout = last_layer.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 106 | let dout = affine2.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 107 | let dout = relu1.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] ... [INFO] [stdout] 112 | db1: affine1.db.clone(), [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = warning: this changes meaning in Rust 2024 [INFO] [stdout] = note: for more information, see issue #123739 [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error: these values and local bindings have significant drop implementation that will have a different drop order from that of Edition 2021 [INFO] [stdout] --> src/two_layer_net.rs:113:25 [INFO] [stdout] | [INFO] [stdout] 91 | let mut affine1 = self.create_affine1(); [INFO] [stdout] | ------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 92 | let mut batch_normalization1 = self.create_batch_normalization1(); [INFO] [stdout] | -------------------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 93 | let mut relu1 = self.create_relu1(); [INFO] [stdout] | ----- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 94 | let mut affine2 = self.create_affine2(); [INFO] [stdout] | ------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 95 | [INFO] [stdout] 96 | let x = affine1.forward(x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 97 | let x = batch_normalization1.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 98 | let x = relu1.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 99 | let x = affine2.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 100 | [INFO] [stdout] 101 | let mut last_layer = SoftmaxWithLossLayer::new(t); [INFO] [stdout] | ---------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] ... [INFO] [stdout] 105 | let dout = last_layer.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 106 | let dout = affine2.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 107 | let dout = relu1.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] ... [INFO] [stdout] 113 | dbatch_aff: batch_normalization1.daff.clone(), [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = warning: this changes meaning in Rust 2024 [INFO] [stdout] = note: for more information, see issue #123739 [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error: these values and local bindings have significant drop implementation that will have a different drop order from that of Edition 2021 [INFO] [stdout] --> src/two_layer_net.rs:114:18 [INFO] [stdout] | [INFO] [stdout] 91 | let mut affine1 = self.create_affine1(); [INFO] [stdout] | ------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 92 | let mut batch_normalization1 = self.create_batch_normalization1(); [INFO] [stdout] | -------------------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 93 | let mut relu1 = self.create_relu1(); [INFO] [stdout] | ----- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 94 | let mut affine2 = self.create_affine2(); [INFO] [stdout] | ------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 95 | [INFO] [stdout] 96 | let x = affine1.forward(x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 97 | let x = batch_normalization1.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 98 | let x = relu1.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 99 | let x = affine2.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 100 | [INFO] [stdout] 101 | let mut last_layer = SoftmaxWithLossLayer::new(t); [INFO] [stdout] | ---------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] ... [INFO] [stdout] 105 | let dout = last_layer.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 106 | let dout = affine2.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 107 | let dout = relu1.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] ... [INFO] [stdout] 114 | dw2: affine2.dw.clone(), [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = warning: this changes meaning in Rust 2024 [INFO] [stdout] = note: for more information, see issue #123739 [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error: these values and local bindings have significant drop implementation that will have a different drop order from that of Edition 2021 [INFO] [stdout] --> src/two_layer_net.rs:115:18 [INFO] [stdout] | [INFO] [stdout] 91 | let mut affine1 = self.create_affine1(); [INFO] [stdout] | ------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 92 | let mut batch_normalization1 = self.create_batch_normalization1(); [INFO] [stdout] | -------------------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 93 | let mut relu1 = self.create_relu1(); [INFO] [stdout] | ----- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 94 | let mut affine2 = self.create_affine2(); [INFO] [stdout] | ------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 95 | [INFO] [stdout] 96 | let x = affine1.forward(x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 97 | let x = batch_normalization1.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 98 | let x = relu1.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 99 | let x = affine2.forward(&x); [INFO] [stdout] | - these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 100 | [INFO] [stdout] 101 | let mut last_layer = SoftmaxWithLossLayer::new(t); [INFO] [stdout] | ---------- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] ... [INFO] [stdout] 105 | let dout = last_layer.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 106 | let dout = affine2.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] 107 | let dout = relu1.backward(&dout); [INFO] [stdout] | ---- these values have significant drop implementation and will observe changes in drop order under Edition 2024 [INFO] [stdout] ... [INFO] [stdout] 115 | db2: affine2.db.clone(), [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = warning: this changes meaning in Rust 2024 [INFO] [stdout] = note: for more information, see issue #123739 [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error: aborting due to 6 previous errors; 20 warnings emitted [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] error: could not compile `zero-deeplearning` (bin "zero-deeplearning" test) due to 7 previous errors; 20 warnings emitted [INFO] running `Command { std: "docker" "inspect" "8461ee927a8ddf79f6a7aa6d8604a4b38bb3345b115a5126a96a29b736daf891", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "8461ee927a8ddf79f6a7aa6d8604a4b38bb3345b115a5126a96a29b736daf891", kill_on_drop: false }` [INFO] [stdout] 8461ee927a8ddf79f6a7aa6d8604a4b38bb3345b115a5126a96a29b736daf891