Nov 28 04:26:32.641 INFO checking helloooooo/learn_deep_learning against try#db42d4dad33013eba11ef37342ad9f614e5652b8 for pr-56074 Nov 28 04:26:32.641 INFO running `"docker" "create" "-v" "/mnt/big/crater/work/local/target-dirs/pr-56074/worker-4/try#db42d4dad33013eba11ef37342ad9f614e5652b8:/target:rw,Z" "-v" "/mnt/big/crater/work/local/test-source/worker-4/pr-56074/try#db42d4dad33013eba11ef37342ad9f614e5652b8:/source:ro,Z" "-v" "/mnt/big/crater/work/local/cargo-home:/cargo-home:ro,Z" "-v" "/mnt/big/crater/work/local/rustup-home:/rustup-home:ro,Z" "-e" "USER_ID=1000" "-e" "SOURCE_DIR=/source" "-e" "USER_ID=1000" "-e" "CMD=cargo +db42d4dad33013eba11ef37342ad9f614e5652b8-alt check --frozen --all --all-targets" "-e" "CARGO_TARGET_DIR=/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/cargo-home" "-e" "RUSTUP_HOME=/rustup-home" "-m" "1536M" "--network" "none" "crater"` Nov 28 04:26:33.247 INFO [stdout] 0e7c1be1c4b01c68104c7e0ea4fc43ea593862447df206e58be38b0083f7f7f5 Nov 28 04:26:33.249 INFO running `"docker" "start" "-a" "0e7c1be1c4b01c68104c7e0ea4fc43ea593862447df206e58be38b0083f7f7f5"` Nov 28 04:26:35.384 INFO [stderr] usermod: no changes Nov 28 04:26:35.706 INFO [stderr] Checking gnuplot v0.0.23 Nov 28 04:26:35.708 INFO [stderr] Checking mnist v0.4.0 Nov 28 04:26:35.711 INFO [stderr] Checking futures-cpupool v0.1.7 Nov 28 04:26:35.718 INFO [stderr] Checking alga v0.5.2 Nov 28 04:26:35.726 INFO [stderr] Compiling openssl v0.9.23 Nov 28 04:26:35.728 INFO [stderr] Checking generic-array v0.8.3 Nov 28 04:26:35.731 INFO [stderr] Checking tokio-proto v0.1.1 Nov 28 04:26:38.624 INFO [stderr] Checking crypto-mac v0.4.0 Nov 28 04:26:38.624 INFO [stderr] Checking digest v0.6.2 Nov 28 04:26:39.809 INFO [stderr] Checking hmac v0.4.2 Nov 28 04:26:39.810 INFO [stderr] Checking sha-1 v0.4.1 Nov 28 04:26:44.661 INFO [stderr] Checking hyper v0.11.9 Nov 28 04:26:45.953 INFO [stderr] Checking native-tls v0.1.4 Nov 28 04:26:46.237 INFO [stderr] Checking nalgebra v0.13.1 Nov 28 04:26:47.162 INFO [stderr] Checking tokio-tls v0.1.3 Nov 28 04:26:55.345 INFO [stderr] Checking hyper-tls v0.1.2 Nov 28 04:26:58.596 INFO [stderr] Checking egg-mode v0.12.0 Nov 28 04:27:25.301 INFO [stderr] Checking test1 v0.1.0 (/source) Nov 28 04:27:26.280 INFO [stderr] warning: unused imports: `RefMut`, `Ref` Nov 28 04:27:26.280 INFO [stderr] --> src/gradient.rs:5:26 Nov 28 04:27:26.280 INFO [stderr] | Nov 28 04:27:26.280 INFO [stderr] 5 | use std::cell::{RefCell, Ref, RefMut}; Nov 28 04:27:26.280 INFO [stderr] | ^^^ ^^^^^^ Nov 28 04:27:26.280 INFO [stderr] | Nov 28 04:27:26.280 INFO [stderr] = note: #[warn(unused_imports)] on by default Nov 28 04:27:26.280 INFO [stderr] Nov 28 04:27:26.280 INFO [stderr] warning: unused import: `std::cell::RefCell` Nov 28 04:27:26.280 INFO [stderr] --> src/nural.rs:5:5 Nov 28 04:27:26.280 INFO [stderr] | Nov 28 04:27:26.280 INFO [stderr] 5 | use std::cell::RefCell; Nov 28 04:27:26.280 INFO [stderr] | ^^^^^^^^^^^^^^^^^^ Nov 28 04:27:26.280 INFO [stderr] Nov 28 04:27:26.496 INFO [stderr] warning: unused variable: `rows` Nov 28 04:27:26.496 INFO [stderr] --> src/main.rs:21:16 Nov 28 04:27:26.496 INFO [stderr] | Nov 28 04:27:26.496 INFO [stderr] 21 | let (size, rows, cols) = (60_000, 28, 28); Nov 28 04:27:26.496 INFO [stderr] | ^^^^ help: consider using `_rows` instead Nov 28 04:27:26.496 INFO [stderr] | Nov 28 04:27:26.496 INFO [stderr] = note: #[warn(unused_variables)] on by default Nov 28 04:27:26.496 INFO [stderr] Nov 28 04:27:26.496 INFO [stderr] warning: unused variable: `cols` Nov 28 04:27:26.496 INFO [stderr] --> src/main.rs:21:22 Nov 28 04:27:26.496 INFO [stderr] | Nov 28 04:27:26.496 INFO [stderr] 21 | let (size, rows, cols) = (60_000, 28, 28); Nov 28 04:27:26.496 INFO [stderr] | ^^^^ help: consider using `_cols` instead Nov 28 04:27:26.496 INFO [stderr] Nov 28 04:27:26.496 INFO [stderr] warning: unused variable: `batch_size` Nov 28 04:27:26.496 INFO [stderr] --> src/main.rs:40:9 Nov 28 04:27:26.496 INFO [stderr] | Nov 28 04:27:26.496 INFO [stderr] 40 | let batch_size = 100; Nov 28 04:27:26.496 INFO [stderr] | ^^^^^^^^^^ help: consider using `_batch_size` instead Nov 28 04:27:26.497 INFO [stderr] Nov 28 04:27:26.524 INFO [stderr] warning: variable does not need to be mutable Nov 28 04:27:26.524 INFO [stderr] --> src/main.rs:33:9 Nov 28 04:27:26.524 INFO [stderr] | Nov 28 04:27:26.524 INFO [stderr] 33 | let mut Two_layer_network = two_layer_net::Two_layer_network { Nov 28 04:27:26.524 INFO [stderr] | ----^^^^^^^^^^^^^^^^^ Nov 28 04:27:26.524 INFO [stderr] | | Nov 28 04:27:26.524 INFO [stderr] | help: remove this `mut` Nov 28 04:27:26.524 INFO [stderr] | Nov 28 04:27:26.524 INFO [stderr] = note: #[warn(unused_mut)] on by default Nov 28 04:27:26.524 INFO [stderr] Nov 28 04:27:26.637 INFO [stderr] warning: function is never used: `mean_squared_error` Nov 28 04:27:26.637 INFO [stderr] --> src/lossfunc.rs:11:1 Nov 28 04:27:26.638 INFO [stderr] | Nov 28 04:27:26.638 INFO [stderr] 11 | pub fn mean_squared_error(y: DMatrix, t: DMatrix) -> f64 { Nov 28 04:27:26.638 INFO [stderr] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Nov 28 04:27:26.638 INFO [stderr] | Nov 28 04:27:26.639 INFO [stderr] = note: #[warn(dead_code)] on by default Nov 28 04:27:26.639 INFO [stderr] Nov 28 04:27:26.639 INFO [stderr] warning: function is never used: `numerical_gradient` Nov 28 04:27:26.640 INFO [stderr] --> src/gradient.rs:8:1 Nov 28 04:27:26.640 INFO [stderr] | Nov 28 04:27:26.640 INFO [stderr] 8 | / pub fn numerical_gradient< Nov 28 04:27:26.640 INFO [stderr] 9 | | F: Fn(&DMatrix, Nov 28 04:27:26.641 INFO [stderr] 10 | | &DMatrix, Nov 28 04:27:26.641 INFO [stderr] 11 | | &DMatrix, Nov 28 04:27:26.641 INFO [stderr] ... | Nov 28 04:27:26.642 INFO [stderr] 38 | | grad Nov 28 04:27:26.642 INFO [stderr] 39 | | } Nov 28 04:27:26.642 INFO [stderr] | |_^ Nov 28 04:27:26.642 INFO [stderr] Nov 28 04:27:26.643 INFO [stderr] warning: function is never used: `function_2` Nov 28 04:27:26.643 INFO [stderr] --> src/gradient.rs:40:1 Nov 28 04:27:26.643 INFO [stderr] | Nov 28 04:27:26.644 INFO [stderr] 40 | pub fn function_2(x: &mut DMatrix) -> f64 { Nov 28 04:27:26.644 INFO [stderr] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Nov 28 04:27:26.644 INFO [stderr] Nov 28 04:27:26.645 INFO [stderr] warning: method is never used: `predict` Nov 28 04:27:26.645 INFO [stderr] --> src/nural.rs:11:5 Nov 28 04:27:26.645 INFO [stderr] | Nov 28 04:27:26.645 INFO [stderr] 11 | pub fn predict(self, x: &DMatrix) -> DMatrix { Nov 28 04:27:26.645 INFO [stderr] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Nov 28 04:27:26.645 INFO [stderr] Nov 28 04:27:26.645 INFO [stderr] warning: method is never used: `loss` Nov 28 04:27:26.645 INFO [stderr] --> src/nural.rs:14:5 Nov 28 04:27:26.645 INFO [stderr] | Nov 28 04:27:26.645 INFO [stderr] 14 | pub fn loss(self, x: &DMatrix, t: &DMatrix) -> f64 { Nov 28 04:27:26.645 INFO [stderr] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Nov 28 04:27:26.645 INFO [stderr] Nov 28 04:27:26.645 INFO [stderr] warning: method is never used: `numerical_gradient` Nov 28 04:27:26.645 INFO [stderr] --> src/two_layer_net.rs:91:5 Nov 28 04:27:26.645 INFO [stderr] | Nov 28 04:27:26.645 INFO [stderr] 91 | pub fn numerical_gradient(&mut self, x: &DMatrix, t: &DMatrix) -> grad { Nov 28 04:27:26.645 INFO [stderr] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Nov 28 04:27:26.645 INFO [stderr] Nov 28 04:27:26.645 INFO [stderr] warning: function is never used: `loss_w` Nov 28 04:27:26.645 INFO [stderr] --> src/two_layer_net.rs:128:1 Nov 28 04:27:26.645 INFO [stderr] | Nov 28 04:27:26.645 INFO [stderr] 128 | / pub fn loss_w( Nov 28 04:27:26.645 INFO [stderr] 129 | | param: &DMatrix, Nov 28 04:27:26.645 INFO [stderr] 130 | | x: &DMatrix, Nov 28 04:27:26.645 INFO [stderr] 131 | | t: &DMatrix, Nov 28 04:27:26.645 INFO [stderr] ... | Nov 28 04:27:26.645 INFO [stderr] 135 | | two.loss(param, x, t, &patern) Nov 28 04:27:26.645 INFO [stderr] 136 | | } Nov 28 04:27:26.645 INFO [stderr] | |_^ Nov 28 04:27:26.645 INFO [stderr] Nov 28 04:27:26.656 INFO [stderr] warning: function `axisZerosum` should have a snake case name such as `axis_zerosum` Nov 28 04:27:26.656 INFO [stderr] --> src/nural.rs:32:1 Nov 28 04:27:26.656 INFO [stderr] | Nov 28 04:27:26.656 INFO [stderr] 32 | / pub fn axisZerosum(x: &DMatrix) -> DMatrix { Nov 28 04:27:26.656 INFO [stderr] 33 | | let zerosum = DMatrix::::from_iterator( Nov 28 04:27:26.656 INFO [stderr] 34 | | 1, Nov 28 04:27:26.656 INFO [stderr] 35 | | x.shape().1, Nov 28 04:27:26.656 INFO [stderr] ... | Nov 28 04:27:26.656 INFO [stderr] 44 | | Nov 28 04:27:26.656 INFO [stderr] 45 | | } Nov 28 04:27:26.656 INFO [stderr] | |_^ Nov 28 04:27:26.656 INFO [stderr] | Nov 28 04:27:26.656 INFO [stderr] = note: #[warn(non_snake_case)] on by default Nov 28 04:27:26.656 INFO [stderr] Nov 28 04:27:26.656 INFO [stderr] warning: function `createVec` should have a snake case name such as `create_vec` Nov 28 04:27:26.656 INFO [stderr] --> src/nural.rs:47:1 Nov 28 04:27:26.656 INFO [stderr] | Nov 28 04:27:26.657 INFO [stderr] 47 | / pub fn createVec(x: usize) -> Vec { Nov 28 04:27:26.657 INFO [stderr] 48 | | let mut vec = Vec::with_capacity(x); Nov 28 04:27:26.657 INFO [stderr] 49 | | for i in 0..x { Nov 28 04:27:26.657 INFO [stderr] 50 | | vec.push(i); Nov 28 04:27:26.657 INFO [stderr] 51 | | } Nov 28 04:27:26.657 INFO [stderr] 52 | | vec Nov 28 04:27:26.657 INFO [stderr] 53 | | } Nov 28 04:27:26.657 INFO [stderr] | |_^ Nov 28 04:27:26.657 INFO [stderr] Nov 28 04:27:26.657 INFO [stderr] warning: type `grad` should have a camel case name such as `Grad` Nov 28 04:27:26.657 INFO [stderr] --> src/two_layer_net.rs:13:1 Nov 28 04:27:26.657 INFO [stderr] | Nov 28 04:27:26.657 INFO [stderr] 13 | / pub struct grad { Nov 28 04:27:26.657 INFO [stderr] 14 | | pub w1: DMatrix, Nov 28 04:27:26.657 INFO [stderr] 15 | | pub b1: DMatrix, Nov 28 04:27:26.657 INFO [stderr] 16 | | pub w2: DMatrix, Nov 28 04:27:26.657 INFO [stderr] 17 | | pub b2: DMatrix, Nov 28 04:27:26.657 INFO [stderr] 18 | | } Nov 28 04:27:26.657 INFO [stderr] | |_^ Nov 28 04:27:26.657 INFO [stderr] | Nov 28 04:27:26.657 INFO [stderr] = note: #[warn(non_camel_case_types)] on by default Nov 28 04:27:26.657 INFO [stderr] Nov 28 04:27:26.657 INFO [stderr] warning: type `Two_layer_network` should have a camel case name such as `TwoLayerNetwork` Nov 28 04:27:26.657 INFO [stderr] --> src/two_layer_net.rs:19:1 Nov 28 04:27:26.657 INFO [stderr] | Nov 28 04:27:26.657 INFO [stderr] 19 | / pub struct Two_layer_network { Nov 28 04:27:26.657 INFO [stderr] 20 | | pub w1: Rc>>, Nov 28 04:27:26.657 INFO [stderr] 21 | | pub b1: Rc>>, Nov 28 04:27:26.657 INFO [stderr] 22 | | pub w2: Rc>>, Nov 28 04:27:26.657 INFO [stderr] 23 | | pub b2: Rc>>, Nov 28 04:27:26.657 INFO [stderr] 24 | | } Nov 28 04:27:26.657 INFO [stderr] | |_^ Nov 28 04:27:26.657 INFO [stderr] Nov 28 04:27:26.657 INFO [stderr] warning: variable `Two_layer_network` should have a snake case name such as `two_layer_network` Nov 28 04:27:26.657 INFO [stderr] --> src/main.rs:33:9 Nov 28 04:27:26.657 INFO [stderr] | Nov 28 04:27:26.657 INFO [stderr] 33 | let mut Two_layer_network = two_layer_net::Two_layer_network { Nov 28 04:27:26.657 INFO [stderr] | ^^^^^^^^^^^^^^^^^^^^^ Nov 28 04:27:26.657 INFO [stderr] Nov 28 04:27:26.696 INFO [stderr] warning: unused imports: `RefMut`, `Ref` Nov 28 04:27:26.696 INFO [stderr] --> src/gradient.rs:5:26 Nov 28 04:27:26.696 INFO [stderr] | Nov 28 04:27:26.696 INFO [stderr] 5 | use std::cell::{RefCell, Ref, RefMut}; Nov 28 04:27:26.696 INFO [stderr] | ^^^ ^^^^^^ Nov 28 04:27:26.696 INFO [stderr] | Nov 28 04:27:26.696 INFO [stderr] = note: #[warn(unused_imports)] on by default Nov 28 04:27:26.696 INFO [stderr] Nov 28 04:27:26.696 INFO [stderr] warning: unused import: `std::cell::RefCell` Nov 28 04:27:26.696 INFO [stderr] --> src/nural.rs:5:5 Nov 28 04:27:26.696 INFO [stderr] | Nov 28 04:27:26.696 INFO [stderr] 5 | use std::cell::RefCell; Nov 28 04:27:26.696 INFO [stderr] | ^^^^^^^^^^^^^^^^^^ Nov 28 04:27:26.696 INFO [stderr] Nov 28 04:27:26.944 INFO [stderr] warning: unused variable: `rows` Nov 28 04:27:26.944 INFO [stderr] --> src/main.rs:21:16 Nov 28 04:27:26.944 INFO [stderr] | Nov 28 04:27:26.945 INFO [stderr] 21 | let (size, rows, cols) = (60_000, 28, 28); Nov 28 04:27:26.945 INFO [stderr] | ^^^^ help: consider using `_rows` instead Nov 28 04:27:26.945 INFO [stderr] | Nov 28 04:27:26.945 INFO [stderr] = note: #[warn(unused_variables)] on by default Nov 28 04:27:26.945 INFO [stderr] Nov 28 04:27:26.945 INFO [stderr] warning: unused variable: `cols` Nov 28 04:27:26.945 INFO [stderr] --> src/main.rs:21:22 Nov 28 04:27:26.945 INFO [stderr] | Nov 28 04:27:26.945 INFO [stderr] 21 | let (size, rows, cols) = (60_000, 28, 28); Nov 28 04:27:26.945 INFO [stderr] | ^^^^ help: consider using `_cols` instead Nov 28 04:27:26.945 INFO [stderr] Nov 28 04:27:26.945 INFO [stderr] warning: unused variable: `batch_size` Nov 28 04:27:26.945 INFO [stderr] --> src/main.rs:40:9 Nov 28 04:27:26.945 INFO [stderr] | Nov 28 04:27:26.945 INFO [stderr] 40 | let batch_size = 100; Nov 28 04:27:26.945 INFO [stderr] | ^^^^^^^^^^ help: consider using `_batch_size` instead Nov 28 04:27:26.945 INFO [stderr] Nov 28 04:27:26.960 INFO [stderr] warning: variable does not need to be mutable Nov 28 04:27:26.960 INFO [stderr] --> src/main.rs:33:9 Nov 28 04:27:26.960 INFO [stderr] | Nov 28 04:27:26.960 INFO [stderr] 33 | let mut Two_layer_network = two_layer_net::Two_layer_network { Nov 28 04:27:26.960 INFO [stderr] | ----^^^^^^^^^^^^^^^^^ Nov 28 04:27:26.960 INFO [stderr] | | Nov 28 04:27:26.960 INFO [stderr] | help: remove this `mut` Nov 28 04:27:26.960 INFO [stderr] | Nov 28 04:27:26.960 INFO [stderr] = note: #[warn(unused_mut)] on by default Nov 28 04:27:26.960 INFO [stderr] Nov 28 04:27:27.008 INFO [stderr] warning: function is never used: `numerical_gradient` Nov 28 04:27:27.008 INFO [stderr] --> src/gradient.rs:8:1 Nov 28 04:27:27.008 INFO [stderr] | Nov 28 04:27:27.008 INFO [stderr] 8 | / pub fn numerical_gradient< Nov 28 04:27:27.008 INFO [stderr] 9 | | F: Fn(&DMatrix, Nov 28 04:27:27.008 INFO [stderr] 10 | | &DMatrix, Nov 28 04:27:27.008 INFO [stderr] 11 | | &DMatrix, Nov 28 04:27:27.008 INFO [stderr] ... | Nov 28 04:27:27.008 INFO [stderr] 38 | | grad Nov 28 04:27:27.008 INFO [stderr] 39 | | } Nov 28 04:27:27.008 INFO [stderr] | |_^ Nov 28 04:27:27.008 INFO [stderr] | Nov 28 04:27:27.008 INFO [stderr] = note: #[warn(dead_code)] on by default Nov 28 04:27:27.009 INFO [stderr] Nov 28 04:27:27.009 INFO [stderr] warning: method is never used: `numerical_gradient` Nov 28 04:27:27.009 INFO [stderr] --> src/two_layer_net.rs:91:5 Nov 28 04:27:27.009 INFO [stderr] | Nov 28 04:27:27.009 INFO [stderr] 91 | pub fn numerical_gradient(&mut self, x: &DMatrix, t: &DMatrix) -> grad { Nov 28 04:27:27.009 INFO [stderr] | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Nov 28 04:27:27.009 INFO [stderr] Nov 28 04:27:27.009 INFO [stderr] warning: function is never used: `loss_w` Nov 28 04:27:27.009 INFO [stderr] --> src/two_layer_net.rs:128:1 Nov 28 04:27:27.009 INFO [stderr] | Nov 28 04:27:27.009 INFO [stderr] 128 | / pub fn loss_w( Nov 28 04:27:27.009 INFO [stderr] 129 | | param: &DMatrix, Nov 28 04:27:27.009 INFO [stderr] 130 | | x: &DMatrix, Nov 28 04:27:27.009 INFO [stderr] 131 | | t: &DMatrix, Nov 28 04:27:27.009 INFO [stderr] ... | Nov 28 04:27:27.009 INFO [stderr] 135 | | two.loss(param, x, t, &patern) Nov 28 04:27:27.009 INFO [stderr] 136 | | } Nov 28 04:27:27.009 INFO [stderr] | |_^ Nov 28 04:27:27.009 INFO [stderr] Nov 28 04:27:27.012 INFO [stderr] warning: function `axisZerosum` should have a snake case name such as `axis_zerosum` Nov 28 04:27:27.012 INFO [stderr] --> src/nural.rs:32:1 Nov 28 04:27:27.012 INFO [stderr] | Nov 28 04:27:27.012 INFO [stderr] 32 | / pub fn axisZerosum(x: &DMatrix) -> DMatrix { Nov 28 04:27:27.012 INFO [stderr] 33 | | let zerosum = DMatrix::::from_iterator( Nov 28 04:27:27.012 INFO [stderr] 34 | | 1, Nov 28 04:27:27.012 INFO [stderr] 35 | | x.shape().1, Nov 28 04:27:27.012 INFO [stderr] ... | Nov 28 04:27:27.012 INFO [stderr] 44 | | Nov 28 04:27:27.012 INFO [stderr] 45 | | } Nov 28 04:27:27.012 INFO [stderr] | |_^ Nov 28 04:27:27.012 INFO [stderr] | Nov 28 04:27:27.012 INFO [stderr] = note: #[warn(non_snake_case)] on by default Nov 28 04:27:27.012 INFO [stderr] Nov 28 04:27:27.012 INFO [stderr] warning: function `createVec` should have a snake case name such as `create_vec` Nov 28 04:27:27.013 INFO [stderr] --> src/nural.rs:47:1 Nov 28 04:27:27.013 INFO [stderr] | Nov 28 04:27:27.013 INFO [stderr] 47 | / pub fn createVec(x: usize) -> Vec { Nov 28 04:27:27.013 INFO [stderr] 48 | | let mut vec = Vec::with_capacity(x); Nov 28 04:27:27.013 INFO [stderr] 49 | | for i in 0..x { Nov 28 04:27:27.013 INFO [stderr] 50 | | vec.push(i); Nov 28 04:27:27.013 INFO [stderr] 51 | | } Nov 28 04:27:27.013 INFO [stderr] 52 | | vec Nov 28 04:27:27.013 INFO [stderr] 53 | | } Nov 28 04:27:27.013 INFO [stderr] | |_^ Nov 28 04:27:27.013 INFO [stderr] Nov 28 04:27:27.013 INFO [stderr] warning: type `grad` should have a camel case name such as `Grad` Nov 28 04:27:27.013 INFO [stderr] --> src/two_layer_net.rs:13:1 Nov 28 04:27:27.013 INFO [stderr] | Nov 28 04:27:27.013 INFO [stderr] 13 | / pub struct grad { Nov 28 04:27:27.013 INFO [stderr] 14 | | pub w1: DMatrix, Nov 28 04:27:27.013 INFO [stderr] 15 | | pub b1: DMatrix, Nov 28 04:27:27.013 INFO [stderr] 16 | | pub w2: DMatrix, Nov 28 04:27:27.013 INFO [stderr] 17 | | pub b2: DMatrix, Nov 28 04:27:27.013 INFO [stderr] 18 | | } Nov 28 04:27:27.013 INFO [stderr] | |_^ Nov 28 04:27:27.013 INFO [stderr] | Nov 28 04:27:27.013 INFO [stderr] = note: #[warn(non_camel_case_types)] on by default Nov 28 04:27:27.013 INFO [stderr] Nov 28 04:27:27.013 INFO [stderr] warning: type `Two_layer_network` should have a camel case name such as `TwoLayerNetwork` Nov 28 04:27:27.013 INFO [stderr] --> src/two_layer_net.rs:19:1 Nov 28 04:27:27.013 INFO [stderr] | Nov 28 04:27:27.013 INFO [stderr] 19 | / pub struct Two_layer_network { Nov 28 04:27:27.013 INFO [stderr] 20 | | pub w1: Rc>>, Nov 28 04:27:27.013 INFO [stderr] 21 | | pub b1: Rc>>, Nov 28 04:27:27.013 INFO [stderr] 22 | | pub w2: Rc>>, Nov 28 04:27:27.013 INFO [stderr] 23 | | pub b2: Rc>>, Nov 28 04:27:27.013 INFO [stderr] 24 | | } Nov 28 04:27:27.013 INFO [stderr] | |_^ Nov 28 04:27:27.013 INFO [stderr] Nov 28 04:27:27.013 INFO [stderr] warning: variable `Two_layer_network` should have a snake case name such as `two_layer_network` Nov 28 04:27:27.013 INFO [stderr] --> src/main.rs:33:9 Nov 28 04:27:27.013 INFO [stderr] | Nov 28 04:27:27.013 INFO [stderr] 33 | let mut Two_layer_network = two_layer_net::Two_layer_network { Nov 28 04:27:27.013 INFO [stderr] | ^^^^^^^^^^^^^^^^^^^^^ Nov 28 04:27:27.013 INFO [stderr] Nov 28 04:27:27.065 INFO [stderr] Finished dev [unoptimized + debuginfo] target(s) in 51.64s Nov 28 04:27:27.088 INFO [stderr] su: No module specific data is present Nov 28 04:27:28.030 INFO running `"docker" "inspect" "0e7c1be1c4b01c68104c7e0ea4fc43ea593862447df206e58be38b0083f7f7f5"` Nov 28 04:27:28.302 INFO running `"docker" "rm" "-f" "0e7c1be1c4b01c68104c7e0ea4fc43ea593862447df206e58be38b0083f7f7f5"` Nov 28 04:27:28.728 INFO [stdout] 0e7c1be1c4b01c68104c7e0ea4fc43ea593862447df206e58be38b0083f7f7f5