[INFO] fetching crate nncombinator 0.3.2... [INFO] checking nncombinator-0.3.2 against master#8c6ce6b91b172f77c795a74bfeaf74b865146b3f for pr-101345 [INFO] extracting crate nncombinator 0.3.2 into /workspace/builds/worker-5/source [INFO] validating manifest of crates.io crate nncombinator 0.3.2 on toolchain 8c6ce6b91b172f77c795a74bfeaf74b865146b3f [INFO] running `Command { std: "/workspace/cargo-home/bin/cargo" "+8c6ce6b91b172f77c795a74bfeaf74b865146b3f" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }` [INFO] started tweaking crates.io crate nncombinator 0.3.2 [INFO] finished tweaking crates.io crate nncombinator 0.3.2 [INFO] tweaked toml for crates.io crate nncombinator 0.3.2 written to /workspace/builds/worker-5/source/Cargo.toml [INFO] running `Command { std: "/workspace/cargo-home/bin/cargo" "+8c6ce6b91b172f77c795a74bfeaf74b865146b3f" "generate-lockfile" "--manifest-path" "Cargo.toml" "-Zno-index-update", kill_on_drop: false }` [INFO] running `Command { std: "/workspace/cargo-home/bin/cargo" "+8c6ce6b91b172f77c795a74bfeaf74b865146b3f" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }` [INFO] [stderr] Downloading crates ... [INFO] [stderr] Downloaded rcublas v0.6.0 [INFO] [stderr] Downloaded cuda-runtime-sys v0.3.0-alpha.1 [INFO] [stderr] Downloaded rcudnn-sys v0.5.0 [INFO] [stderr] Downloaded cuda-config v0.1.0 [INFO] [stderr] Downloaded rcublas-sys v0.5.0 [INFO] [stderr] Downloaded rcudnn v1.8.0 [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-5/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-5/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:824c28ce115f6b999260af6986f3384c158e782489540e41c0b698ca1d9fd927" "/opt/rustwide/cargo-home/bin/cargo" "+8c6ce6b91b172f77c795a74bfeaf74b865146b3f" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }` [INFO] [stdout] e377160fd4dea09c8f5666ca8e5114fc9e870450e54c940c5269009bd1e810e3 [INFO] running `Command { std: "docker" "start" "-a" "e377160fd4dea09c8f5666ca8e5114fc9e870450e54c940c5269009bd1e810e3", kill_on_drop: false }` [INFO] running `Command { std: "docker" "inspect" "e377160fd4dea09c8f5666ca8e5114fc9e870450e54c940c5269009bd1e810e3", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "e377160fd4dea09c8f5666ca8e5114fc9e870450e54c940c5269009bd1e810e3", kill_on_drop: false }` [INFO] [stdout] e377160fd4dea09c8f5666ca8e5114fc9e870450e54c940c5269009bd1e810e3 [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-5/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-5/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:824c28ce115f6b999260af6986f3384c158e782489540e41c0b698ca1d9fd927" "/opt/rustwide/cargo-home/bin/cargo" "+8c6ce6b91b172f77c795a74bfeaf74b865146b3f" "check" "--frozen" "--all" "--all-targets" "--message-format=json", kill_on_drop: false }` [INFO] [stdout] bdc2744c65d5cafb586f613743af5180380eed67b079bc4211caa3b9b023e17d [INFO] running `Command { std: "docker" "start" "-a" "bdc2744c65d5cafb586f613743af5180380eed67b079bc4211caa3b9b023e17d", kill_on_drop: false }` [INFO] [stderr] Compiling autocfg v1.1.0 [INFO] [stderr] Compiling libc v0.2.132 [INFO] [stderr] Checking cfg-if v1.0.0 [INFO] [stderr] Compiling libm v0.2.5 [INFO] [stderr] Compiling proc-macro2 v1.0.43 [INFO] [stderr] Compiling unicode-ident v1.0.3 [INFO] [stderr] Compiling quote v1.0.21 [INFO] [stderr] Compiling pkg-config v0.3.25 [INFO] [stderr] Compiling syn v1.0.99 [INFO] [stderr] Compiling crossbeam-utils v0.8.11 [INFO] [stderr] Checking once_cell v1.14.0 [INFO] [stderr] Compiling glob v0.3.0 [INFO] [stderr] Checking scopeguard v1.1.0 [INFO] [stderr] Checking lazy_static v1.4.0 [INFO] [stderr] Compiling rayon-core v1.9.3 [INFO] [stderr] Checking ppv-lite86 v0.2.16 [INFO] [stderr] Compiling memchr v2.5.0 [INFO] [stderr] Compiling log v0.4.17 [INFO] [stderr] Compiling serde v1.0.144 [INFO] [stderr] Compiling typenum v1.15.0 [INFO] [stderr] Checking rawpointer v0.2.1 [INFO] [stderr] Compiling paste v1.0.9 [INFO] [stderr] Compiling cc v1.0.73 [INFO] [stderr] Checking regex-automata v0.1.10 [INFO] [stderr] Checking either v1.8.0 [INFO] [stderr] Checking ryu v1.0.11 [INFO] [stderr] Checking itoa v0.4.8 [INFO] [stderr] Checking matrixmultiply v0.3.2 [INFO] [stderr] Compiling cuda-config v0.1.0 [INFO] [stderr] Compiling num-traits v0.2.15 [INFO] [stderr] Compiling num-integer v0.1.45 [INFO] [stderr] Compiling num-bigint v0.4.3 [INFO] [stderr] Compiling memoffset v0.6.5 [INFO] [stderr] Compiling num-rational v0.4.1 [INFO] [stderr] Compiling crossbeam-epoch v0.9.10 [INFO] [stderr] Compiling num-iter v0.1.43 [INFO] [stderr] Compiling rayon v1.5.3 [INFO] [stderr] Checking crossbeam-channel v0.5.6 [INFO] [stderr] Compiling rcudnn-sys v0.5.0 [INFO] [stderr] Compiling rcublas-sys v0.5.0 [INFO] [stderr] Checking csv-core v0.1.10 [INFO] [stderr] Compiling nncombinator v0.3.2 (/opt/rustwide/workdir) [INFO] [stderr] Checking crossbeam-deque v0.8.2 [INFO] [stderr] Checking getrandom v0.2.7 [INFO] [stderr] Checking num_cpus v1.13.1 [INFO] [stderr] Checking rand_core v0.6.3 [INFO] [stderr] Checking rand_chacha v0.3.1 [INFO] [stderr] Checking rand_xorshift v0.3.0 [INFO] [stderr] Checking num-complex v0.4.2 [INFO] [stderr] Checking approx v0.5.1 [INFO] [stderr] Checking rand v0.8.5 [INFO] [stderr] Compiling cuda-runtime-sys v0.3.0-alpha.1 [INFO] [stderr] Checking simba v0.5.1 [INFO] [stderr] Checking rand_distr v0.4.3 [INFO] [stderr] Checking num v0.4.0 [INFO] [stderr] Checking bstr v0.2.17 [INFO] [stderr] Checking csv v1.1.6 [INFO] [stderr] Compiling thiserror-impl v1.0.33 [INFO] [stderr] Compiling nalgebra-macros v0.1.0 [INFO] [stderr] Checking nalgebra v0.27.1 [INFO] [stderr] Checking thiserror v1.0.33 [INFO] [stderr] Checking rcublas v0.6.0 [INFO] [stderr] Checking rcudnn v1.8.0 [INFO] [stdout] warning: unused import: `KernelArgs` [INFO] [stdout] --> src/activation.rs:7:50 [INFO] [stdout] | [INFO] [stdout] 7 | use crate::cuda::{CudaPtr, DataTypeInfo, Kernel, KernelArgs, Memory}; [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `std::fmt::Debug` [INFO] [stdout] --> src/device.rs:1:5 [INFO] [stdout] | [INFO] [stdout] 1 | use std::fmt::Debug; [INFO] [stdout] | ^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused imports: `Cudnn`, `TensorDescriptor` [INFO] [stdout] --> src/device.rs:9:14 [INFO] [stdout] | [INFO] [stdout] 9 | use rcudnn::{Cudnn, TensorDescriptor}; [INFO] [stdout] | ^^^^^ ^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `rcudnn::utils::DataType` [INFO] [stdout] --> src/device.rs:10:5 [INFO] [stdout] | [INFO] [stdout] 10 | use rcudnn::utils::DataType; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `l` [INFO] [stdout] --> src/activation.rs:260:53 [INFO] [stdout] | [INFO] [stdout] 260 | fn is_canonical_link>(&self, l: &L) -> bool { [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_l` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_variables)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `l` [INFO] [stdout] --> src/activation.rs:347:53 [INFO] [stdout] | [INFO] [stdout] 347 | fn is_canonical_link>(&self, l: &L) -> bool { [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_l` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `l` [INFO] [stdout] --> src/activation.rs:433:53 [INFO] [stdout] | [INFO] [stdout] 433 | fn is_canonical_link>(&self, l: &L) -> bool { [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_l` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `o` [INFO] [stdout] --> src/device.rs:363:63 [INFO] [stdout] | [INFO] [stdout] 363 | fn batch_loss_linear_by_activaton(&self, o:&Vec>, loss:&Vec>, u:&Vec>, activation:&A) [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_o` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `loss` [INFO] [stdout] --> src/device.rs:363:83 [INFO] [stdout] | [INFO] [stdout] 363 | fn batch_loss_linear_by_activaton(&self, o:&Vec>, loss:&Vec>, u:&Vec>, activation:&A) [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_loss` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `u` [INFO] [stdout] --> src/device.rs:363:106 [INFO] [stdout] | [INFO] [stdout] 363 | fn batch_loss_linear_by_activaton(&self, o:&Vec>, loss:&Vec>, u:&Vec>, activation:&A) [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_u` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `activation` [INFO] [stdout] --> src/device.rs:363:126 [INFO] [stdout] | [INFO] [stdout] 363 | fn batch_loss_linear_by_activaton(&self, o:&Vec>, loss:&Vec>, u:&Vec>, activation:&A) [INFO] [stdout] | ^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_activation` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `o` [INFO] [stdout] --> src/device.rs:497:63 [INFO] [stdout] | [INFO] [stdout] 497 | fn batch_loss_linear_by_activaton(&self, o:&Vec>, loss:&Vec>, u:&Vec>, activation:&A) [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_o` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `loss` [INFO] [stdout] --> src/device.rs:497:83 [INFO] [stdout] | [INFO] [stdout] 497 | fn batch_loss_linear_by_activaton(&self, o:&Vec>, loss:&Vec>, u:&Vec>, activation:&A) [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_loss` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `u` [INFO] [stdout] --> src/device.rs:497:106 [INFO] [stdout] | [INFO] [stdout] 497 | fn batch_loss_linear_by_activaton(&self, o:&Vec>, loss:&Vec>, u:&Vec>, activation:&A) [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_u` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `activation` [INFO] [stdout] --> src/device.rs:497:126 [INFO] [stdout] | [INFO] [stdout] 497 | fn batch_loss_linear_by_activaton(&self, o:&Vec>, loss:&Vec>, u:&Vec>, activation:&A) [INFO] [stdout] | ^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_activation` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `input` [INFO] [stdout] --> src/layer.rs:374:29 [INFO] [stdout] | [INFO] [stdout] 374 | fn batch_forward(&self, input: Self::BatchInput) -> Result { [INFO] [stdout] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_input` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `input` [INFO] [stdout] --> src/layer.rs:414:31 [INFO] [stdout] | [INFO] [stdout] 414 | fn batch_pre_train(&self, input: Self::BatchInput) -> Result { [INFO] [stdout] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_input` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `input` [INFO] [stdout] --> src/layer.rs:1108:29 [INFO] [stdout] | [INFO] [stdout] 1108 | fn batch_forward(&self, input: Self::BatchInput) -> Result { [INFO] [stdout] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_input` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `input` [INFO] [stdout] --> src/layer.rs:1144:31 [INFO] [stdout] | [INFO] [stdout] 1144 | fn batch_pre_train(&self, input: Self::BatchInput) -> Result { [INFO] [stdout] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_input` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `input` [INFO] [stdout] --> src/layer.rs:1213:72 [INFO] [stdout] | [INFO] [stdout] 1213 | ...: LossFunction>(&mut self, input: Self::BatchLossInput, stack: Self::BatchOutStack, optimizer: &mut OP, lossf: &L) -> Result<(), Tr... [INFO] [stdout] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_input` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `stack` [INFO] [stdout] --> src/layer.rs:1213:101 [INFO] [stdout] | [INFO] [stdout] 1213 | ... input: Self::BatchLossInput, stack: Self::BatchOutStack, optimizer: &mut OP, lossf: &L) -> Result<(), TrainingError> { [INFO] [stdout] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_stack` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `optimizer` [INFO] [stdout] --> src/layer.rs:1213:129 [INFO] [stdout] | [INFO] [stdout] 1213 | ...ack: Self::BatchOutStack, optimizer: &mut OP, lossf: &L) -> Result<(), TrainingError> { [INFO] [stdout] | ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_optimizer` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `lossf` [INFO] [stdout] --> src/layer.rs:1213:149 [INFO] [stdout] | [INFO] [stdout] 1213 | ...OutStack, optimizer: &mut OP, lossf: &L) -> Result<(), TrainingError> { [INFO] [stdout] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_lossf` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variable does not need to be mutable [INFO] [stdout] --> src/activation.rs:500:13 [INFO] [stdout] | [INFO] [stdout] 500 | let mut alpha = CudaPtr::try_from(alpha)?; [INFO] [stdout] | ----^^^^^ [INFO] [stdout] | | [INFO] [stdout] | help: remove this `mut` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_mut)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variable does not need to be mutable [INFO] [stdout] --> src/activation.rs:501:13 [INFO] [stdout] | [INFO] [stdout] 501 | let mut sum = CudaPtr::try_from(sum)?; [INFO] [stdout] | ----^^^ [INFO] [stdout] | | [INFO] [stdout] | help: remove this `mut` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `swish_forward_double` is never used [INFO] [stdout] --> src/cuda/kernel/activation.rs:19:8 [INFO] [stdout] | [INFO] [stdout] 19 | fn swish_forward_double(input_output: *mut f64, len: size_t, units_len: size_t) -> c_void; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(dead_code)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `softmax_preprocessing_double` is never used [INFO] [stdout] --> src/cuda/kernel/activation.rs:27:8 [INFO] [stdout] | [INFO] [stdout] 27 | fn softmax_preprocessing_double(input: *const f64, len: size_t, batch_len: size_t, alpha: *mut f64, sum: *mut f64) -> c_void; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: 27 warnings emitted [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] Checking statrs v0.15.0 [INFO] [stdout] warning: unused import: `KernelArgs` [INFO] [stdout] --> src/activation.rs:7:50 [INFO] [stdout] | [INFO] [stdout] 7 | use crate::cuda::{CudaPtr, DataTypeInfo, Kernel, KernelArgs, Memory}; [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `std::fmt::Debug` [INFO] [stdout] --> src/device.rs:1:5 [INFO] [stdout] | [INFO] [stdout] 1 | use std::fmt::Debug; [INFO] [stdout] | ^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused imports: `Cudnn`, `TensorDescriptor` [INFO] [stdout] --> src/device.rs:9:14 [INFO] [stdout] | [INFO] [stdout] 9 | use rcudnn::{Cudnn, TensorDescriptor}; [INFO] [stdout] | ^^^^^ ^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `rcudnn::utils::DataType` [INFO] [stdout] --> src/device.rs:10:5 [INFO] [stdout] | [INFO] [stdout] 10 | use rcudnn::utils::DataType; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `l` [INFO] [stdout] --> src/activation.rs:260:53 [INFO] [stdout] | [INFO] [stdout] 260 | fn is_canonical_link>(&self, l: &L) -> bool { [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_l` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_variables)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `l` [INFO] [stdout] --> src/activation.rs:347:53 [INFO] [stdout] | [INFO] [stdout] 347 | fn is_canonical_link>(&self, l: &L) -> bool { [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_l` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `l` [INFO] [stdout] --> src/activation.rs:433:53 [INFO] [stdout] | [INFO] [stdout] 433 | fn is_canonical_link>(&self, l: &L) -> bool { [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_l` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `o` [INFO] [stdout] --> src/device.rs:363:63 [INFO] [stdout] | [INFO] [stdout] 363 | fn batch_loss_linear_by_activaton(&self, o:&Vec>, loss:&Vec>, u:&Vec>, activation:&A) [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_o` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `loss` [INFO] [stdout] --> src/device.rs:363:83 [INFO] [stdout] | [INFO] [stdout] 363 | fn batch_loss_linear_by_activaton(&self, o:&Vec>, loss:&Vec>, u:&Vec>, activation:&A) [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_loss` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `u` [INFO] [stdout] --> src/device.rs:363:106 [INFO] [stdout] | [INFO] [stdout] 363 | fn batch_loss_linear_by_activaton(&self, o:&Vec>, loss:&Vec>, u:&Vec>, activation:&A) [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_u` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `activation` [INFO] [stdout] --> src/device.rs:363:126 [INFO] [stdout] | [INFO] [stdout] 363 | fn batch_loss_linear_by_activaton(&self, o:&Vec>, loss:&Vec>, u:&Vec>, activation:&A) [INFO] [stdout] | ^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_activation` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `o` [INFO] [stdout] --> src/device.rs:497:63 [INFO] [stdout] | [INFO] [stdout] 497 | fn batch_loss_linear_by_activaton(&self, o:&Vec>, loss:&Vec>, u:&Vec>, activation:&A) [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_o` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `loss` [INFO] [stdout] --> src/device.rs:497:83 [INFO] [stdout] | [INFO] [stdout] 497 | fn batch_loss_linear_by_activaton(&self, o:&Vec>, loss:&Vec>, u:&Vec>, activation:&A) [INFO] [stdout] | ^^^^ help: if this is intentional, prefix it with an underscore: `_loss` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `u` [INFO] [stdout] --> src/device.rs:497:106 [INFO] [stdout] | [INFO] [stdout] 497 | fn batch_loss_linear_by_activaton(&self, o:&Vec>, loss:&Vec>, u:&Vec>, activation:&A) [INFO] [stdout] | ^ help: if this is intentional, prefix it with an underscore: `_u` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `activation` [INFO] [stdout] --> src/device.rs:497:126 [INFO] [stdout] | [INFO] [stdout] 497 | fn batch_loss_linear_by_activaton(&self, o:&Vec>, loss:&Vec>, u:&Vec>, activation:&A) [INFO] [stdout] | ^^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_activation` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `input` [INFO] [stdout] --> src/layer.rs:374:29 [INFO] [stdout] | [INFO] [stdout] 374 | fn batch_forward(&self, input: Self::BatchInput) -> Result { [INFO] [stdout] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_input` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `input` [INFO] [stdout] --> src/layer.rs:414:31 [INFO] [stdout] | [INFO] [stdout] 414 | fn batch_pre_train(&self, input: Self::BatchInput) -> Result { [INFO] [stdout] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_input` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `input` [INFO] [stdout] --> src/layer.rs:1108:29 [INFO] [stdout] | [INFO] [stdout] 1108 | fn batch_forward(&self, input: Self::BatchInput) -> Result { [INFO] [stdout] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_input` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `input` [INFO] [stdout] --> src/layer.rs:1144:31 [INFO] [stdout] | [INFO] [stdout] 1144 | fn batch_pre_train(&self, input: Self::BatchInput) -> Result { [INFO] [stdout] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_input` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `input` [INFO] [stdout] --> src/layer.rs:1213:72 [INFO] [stdout] | [INFO] [stdout] 1213 | ...: LossFunction>(&mut self, input: Self::BatchLossInput, stack: Self::BatchOutStack, optimizer: &mut OP, lossf: &L) -> Result<(), Tr... [INFO] [stdout] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_input` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `stack` [INFO] [stdout] --> src/layer.rs:1213:101 [INFO] [stdout] | [INFO] [stdout] 1213 | ... input: Self::BatchLossInput, stack: Self::BatchOutStack, optimizer: &mut OP, lossf: &L) -> Result<(), TrainingError> { [INFO] [stdout] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_stack` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `optimizer` [INFO] [stdout] --> src/layer.rs:1213:129 [INFO] [stdout] | [INFO] [stdout] 1213 | ...ack: Self::BatchOutStack, optimizer: &mut OP, lossf: &L) -> Result<(), TrainingError> { [INFO] [stdout] | ^^^^^^^^^ help: if this is intentional, prefix it with an underscore: `_optimizer` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused variable: `lossf` [INFO] [stdout] --> src/layer.rs:1213:149 [INFO] [stdout] | [INFO] [stdout] 1213 | ...OutStack, optimizer: &mut OP, lossf: &L) -> Result<(), TrainingError> { [INFO] [stdout] | ^^^^^ help: if this is intentional, prefix it with an underscore: `_lossf` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variable does not need to be mutable [INFO] [stdout] --> src/activation.rs:500:13 [INFO] [stdout] | [INFO] [stdout] 500 | let mut alpha = CudaPtr::try_from(alpha)?; [INFO] [stdout] | ----^^^^^ [INFO] [stdout] | | [INFO] [stdout] | help: remove this `mut` [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_mut)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variable does not need to be mutable [INFO] [stdout] --> src/activation.rs:501:13 [INFO] [stdout] | [INFO] [stdout] 501 | let mut sum = CudaPtr::try_from(sum)?; [INFO] [stdout] | ----^^^ [INFO] [stdout] | | [INFO] [stdout] | help: remove this `mut` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `swish_forward_double` is never used [INFO] [stdout] --> src/cuda/kernel/activation.rs:19:8 [INFO] [stdout] | [INFO] [stdout] 19 | fn swish_forward_double(input_output: *mut f64, len: size_t, units_len: size_t) -> c_void; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(dead_code)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `softmax_preprocessing_double` is never used [INFO] [stdout] --> src/cuda/kernel/activation.rs:27:8 [INFO] [stdout] | [INFO] [stdout] 27 | fn softmax_preprocessing_double(input: *const f64, len: size_t, batch_len: size_t, alpha: *mut f64, sum: *mut f64) -> c_void; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: 27 warnings emitted [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] Finished dev [unoptimized + debuginfo] target(s) in 24.68s [INFO] running `Command { std: "docker" "inspect" "bdc2744c65d5cafb586f613743af5180380eed67b079bc4211caa3b9b023e17d", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "bdc2744c65d5cafb586f613743af5180380eed67b079bc4211caa3b9b023e17d", kill_on_drop: false }` [INFO] [stdout] bdc2744c65d5cafb586f613743af5180380eed67b079bc4211caa3b9b023e17d