Nov 01 04:11:56.291 INFO checking helloooooo/learn_deep_learning against try#d7816399d50d407bbce1efe9dbadf8faaaf568fc for pr-55538 Nov 01 04:11:56.292 INFO running `"docker" "create" "-v" "/mnt/big/crater/work/local/target-dirs/pr-55538/worker-4/try#d7816399d50d407bbce1efe9dbadf8faaaf568fc:/target:rw,Z" "-v" "/mnt/big/crater/work/local/test-source/worker-4/pr-55538/try#d7816399d50d407bbce1efe9dbadf8faaaf568fc:/source:ro,Z" "-v" "/mnt/big/crater/work/local/cargo-home:/cargo-home:ro,Z" "-v" "/mnt/big/crater/work/local/rustup-home:/rustup-home:ro,Z" "-e" "USER_ID=1000" "-e" "SOURCE_DIR=/source" "-e" "USER_ID=1000" "-e" "CMD=cargo +d7816399d50d407bbce1efe9dbadf8faaaf568fc-alt check --frozen --all --all-targets" "-e" "CARGO_TARGET_DIR=/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/cargo-home" "-e" "RUSTUP_HOME=/rustup-home" "-m" "1536M" "--network" "none" "crater"` Nov 01 04:11:56.757 INFO blam! 4131e103227979b3f1eba89abfc82e7cdb552d71d87f730dd5b2e604acd4db9e Nov 01 04:11:56.758 INFO running `"docker" "start" "-a" "4131e103227979b3f1eba89abfc82e7cdb552d71d87f730dd5b2e604acd4db9e"` Nov 01 04:11:57.900 INFO kablam! usermod: no changes Nov 01 04:11:58.162 INFO kablam! Checking gnuplot v0.0.23 Nov 01 04:11:58.181 INFO kablam! Checking mnist v0.4.0 Nov 01 04:11:58.193 INFO kablam! Checking alga v0.5.2 Nov 01 04:11:58.196 INFO kablam! Checking generic-array v0.8.3 Nov 01 04:11:58.197 INFO kablam! Checking native-tls v0.1.4 Nov 01 04:11:59.684 INFO kablam! Checking digest v0.6.2 Nov 01 04:11:59.684 INFO kablam! Checking crypto-mac v0.4.0 Nov 01 04:12:00.512 INFO kablam! Checking sha-1 v0.4.1 Nov 01 04:12:00.604 INFO kablam! Checking hmac v0.4.2 Nov 01 04:12:01.690 INFO kablam! Checking tokio-tls v0.1.3 Nov 01 04:12:04.077 INFO kablam! Checking hyper-tls v0.1.2 Nov 01 04:12:07.189 INFO kablam! Checking egg-mode v0.12.0 Nov 01 04:12:09.629 INFO kablam! Checking nalgebra v0.13.1 Nov 01 04:12:43.276 INFO kablam! Checking test1 v0.1.0 (/source) Nov 01 04:12:44.032 INFO kablam! warning: unused imports: `RefMut`, `Ref` Nov 01 04:12:44.032 INFO kablam! --> src/gradient.rs:5:26 Nov 01 04:12:44.032 INFO kablam! | Nov 01 04:12:44.032 INFO kablam! 5 | use std::cell::{RefCell, Ref, RefMut}; Nov 01 04:12:44.032 INFO kablam! | ^^^ ^^^^^^ Nov 01 04:12:44.032 INFO kablam! | Nov 01 04:12:44.032 INFO kablam! = note: #[warn(unused_imports)] on by default Nov 01 04:12:44.032 INFO kablam! Nov 01 04:12:44.032 INFO kablam! warning: unused import: `std::cell::RefCell` Nov 01 04:12:44.032 INFO kablam! --> src/nural.rs:5:5 Nov 01 04:12:44.032 INFO kablam! | Nov 01 04:12:44.032 INFO kablam! 5 | use std::cell::RefCell; Nov 01 04:12:44.032 INFO kablam! | ^^^^^^^^^^^^^^^^^^ Nov 01 04:12:44.032 INFO kablam! Nov 01 04:12:44.324 INFO kablam! warning: unused variable: `rows` Nov 01 04:12:44.324 INFO kablam! --> src/main.rs:21:16 Nov 01 04:12:44.324 INFO kablam! | Nov 01 04:12:44.324 INFO kablam! 21 | let (size, rows, cols) = (60_000, 28, 28); Nov 01 04:12:44.324 INFO kablam! | ^^^^ help: consider using `_rows` instead Nov 01 04:12:44.324 INFO kablam! | Nov 01 04:12:44.324 INFO kablam! = note: #[warn(unused_variables)] on by default Nov 01 04:12:44.324 INFO kablam! Nov 01 04:12:44.324 INFO kablam! warning: unused variable: `cols` Nov 01 04:12:44.324 INFO kablam! --> src/main.rs:21:22 Nov 01 04:12:44.324 INFO kablam! | Nov 01 04:12:44.324 INFO kablam! 21 | let (size, rows, cols) = (60_000, 28, 28); Nov 01 04:12:44.324 INFO kablam! | ^^^^ help: consider using `_cols` instead Nov 01 04:12:44.325 INFO kablam! Nov 01 04:12:44.325 INFO kablam! warning: unused variable: `batch_size` Nov 01 04:12:44.325 INFO kablam! --> src/main.rs:40:9 Nov 01 04:12:44.325 INFO kablam! | Nov 01 04:12:44.325 INFO kablam! 40 | let batch_size = 100; Nov 01 04:12:44.325 INFO kablam! | ^^^^^^^^^^ help: consider using `_batch_size` instead Nov 01 04:12:44.325 INFO kablam! Nov 01 04:12:44.348 INFO kablam! warning: variable does not need to be mutable Nov 01 04:12:44.348 INFO kablam! --> src/main.rs:33:9 Nov 01 04:12:44.348 INFO kablam! | Nov 01 04:12:44.348 INFO kablam! 33 | let mut Two_layer_network = two_layer_net::Two_layer_network { Nov 01 04:12:44.348 INFO kablam! | ----^^^^^^^^^^^^^^^^^ Nov 01 04:12:44.348 INFO kablam! | | Nov 01 04:12:44.348 INFO kablam! | help: remove this `mut` Nov 01 04:12:44.348 INFO kablam! | Nov 01 04:12:44.348 INFO kablam! = note: #[warn(unused_mut)] on by default Nov 01 04:12:44.348 INFO kablam! Nov 01 04:12:44.392 INFO kablam! warning: function is never used: `numerical_gradient` Nov 01 04:12:44.392 INFO kablam! --> src/gradient.rs:8:1 Nov 01 04:12:44.392 INFO kablam! | Nov 01 04:12:44.392 INFO kablam! 8 | / pub fn numerical_gradient< Nov 01 04:12:44.392 INFO kablam! 9 | | F: Fn(&DMatrix, Nov 01 04:12:44.392 INFO kablam! 10 | | &DMatrix, Nov 01 04:12:44.392 INFO kablam! 11 | | &DMatrix, Nov 01 04:12:44.392 INFO kablam! ... | Nov 01 04:12:44.392 INFO kablam! 38 | | grad Nov 01 04:12:44.392 INFO kablam! 39 | | } Nov 01 04:12:44.392 INFO kablam! | |_^ Nov 01 04:12:44.392 INFO kablam! | Nov 01 04:12:44.392 INFO kablam! = note: #[warn(dead_code)] on by default Nov 01 04:12:44.392 INFO kablam! Nov 01 04:12:44.392 INFO kablam! warning: method is never used: `numerical_gradient` Nov 01 04:12:44.392 INFO kablam! --> src/two_layer_net.rs:91:5 Nov 01 04:12:44.392 INFO kablam! | Nov 01 04:12:44.392 INFO kablam! 91 | pub fn numerical_gradient(&mut self, x: &DMatrix, t: &DMatrix) -> grad { Nov 01 04:12:44.392 INFO kablam! | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Nov 01 04:12:44.392 INFO kablam! Nov 01 04:12:44.392 INFO kablam! warning: function is never used: `loss_w` Nov 01 04:12:44.392 INFO kablam! --> src/two_layer_net.rs:128:1 Nov 01 04:12:44.393 INFO kablam! | Nov 01 04:12:44.393 INFO kablam! 128 | / pub fn loss_w( Nov 01 04:12:44.393 INFO kablam! 129 | | param: &DMatrix, Nov 01 04:12:44.393 INFO kablam! 130 | | x: &DMatrix, Nov 01 04:12:44.393 INFO kablam! 131 | | t: &DMatrix, Nov 01 04:12:44.393 INFO kablam! ... | Nov 01 04:12:44.393 INFO kablam! 135 | | two.loss(param, x, t, &patern) Nov 01 04:12:44.393 INFO kablam! 136 | | } Nov 01 04:12:44.393 INFO kablam! | |_^ Nov 01 04:12:44.393 INFO kablam! Nov 01 04:12:44.393 INFO kablam! warning: function `axisZerosum` should have a snake case name such as `axis_zerosum` Nov 01 04:12:44.393 INFO kablam! --> src/nural.rs:32:1 Nov 01 04:12:44.393 INFO kablam! | Nov 01 04:12:44.393 INFO kablam! 32 | / pub fn axisZerosum(x: &DMatrix) -> DMatrix { Nov 01 04:12:44.393 INFO kablam! 33 | | let zerosum = DMatrix::::from_iterator( Nov 01 04:12:44.393 INFO kablam! 34 | | 1, Nov 01 04:12:44.393 INFO kablam! 35 | | x.shape().1, Nov 01 04:12:44.393 INFO kablam! ... | Nov 01 04:12:44.393 INFO kablam! 44 | | Nov 01 04:12:44.393 INFO kablam! 45 | | } Nov 01 04:12:44.393 INFO kablam! | |_^ Nov 01 04:12:44.393 INFO kablam! | Nov 01 04:12:44.393 INFO kablam! = note: #[warn(non_snake_case)] on by default Nov 01 04:12:44.393 INFO kablam! Nov 01 04:12:44.393 INFO kablam! warning: function `createVec` should have a snake case name such as `create_vec` Nov 01 04:12:44.393 INFO kablam! --> src/nural.rs:47:1 Nov 01 04:12:44.393 INFO kablam! | Nov 01 04:12:44.393 INFO kablam! 47 | / pub fn createVec(x: usize) -> Vec { Nov 01 04:12:44.393 INFO kablam! 48 | | let mut vec = Vec::with_capacity(x); Nov 01 04:12:44.393 INFO kablam! 49 | | for i in 0..x { Nov 01 04:12:44.393 INFO kablam! 50 | | vec.push(i); Nov 01 04:12:44.393 INFO kablam! 51 | | } Nov 01 04:12:44.393 INFO kablam! 52 | | vec Nov 01 04:12:44.393 INFO kablam! 53 | | } Nov 01 04:12:44.393 INFO kablam! | |_^ Nov 01 04:12:44.393 INFO kablam! Nov 01 04:12:44.395 INFO kablam! warning: type `grad` should have a camel case name such as `Grad` Nov 01 04:12:44.395 INFO kablam! --> src/two_layer_net.rs:13:1 Nov 01 04:12:44.395 INFO kablam! | Nov 01 04:12:44.395 INFO kablam! 13 | / pub struct grad { Nov 01 04:12:44.395 INFO kablam! 14 | | pub w1: DMatrix, Nov 01 04:12:44.395 INFO kablam! 15 | | pub b1: DMatrix, Nov 01 04:12:44.395 INFO kablam! 16 | | pub w2: DMatrix, Nov 01 04:12:44.395 INFO kablam! 17 | | pub b2: DMatrix, Nov 01 04:12:44.395 INFO kablam! 18 | | } Nov 01 04:12:44.395 INFO kablam! | |_^ Nov 01 04:12:44.395 INFO kablam! | Nov 01 04:12:44.395 INFO kablam! = note: #[warn(non_camel_case_types)] on by default Nov 01 04:12:44.395 INFO kablam! Nov 01 04:12:44.395 INFO kablam! warning: type `Two_layer_network` should have a camel case name such as `TwoLayerNetwork` Nov 01 04:12:44.395 INFO kablam! --> src/two_layer_net.rs:19:1 Nov 01 04:12:44.395 INFO kablam! | Nov 01 04:12:44.395 INFO kablam! 19 | / pub struct Two_layer_network { Nov 01 04:12:44.395 INFO kablam! 20 | | pub w1: Rc>>, Nov 01 04:12:44.395 INFO kablam! 21 | | pub b1: Rc>>, Nov 01 04:12:44.395 INFO kablam! 22 | | pub w2: Rc>>, Nov 01 04:12:44.395 INFO kablam! 23 | | pub b2: Rc>>, Nov 01 04:12:44.395 INFO kablam! 24 | | } Nov 01 04:12:44.395 INFO kablam! | |_^ Nov 01 04:12:44.395 INFO kablam! Nov 01 04:12:44.395 INFO kablam! warning: variable `Two_layer_network` should have a snake case name such as `two_layer_network` Nov 01 04:12:44.395 INFO kablam! --> src/main.rs:33:9 Nov 01 04:12:44.395 INFO kablam! | Nov 01 04:12:44.395 INFO kablam! 33 | let mut Two_layer_network = two_layer_net::Two_layer_network { Nov 01 04:12:44.395 INFO kablam! | ^^^^^^^^^^^^^^^^^^^^^ Nov 01 04:12:44.395 INFO kablam! Nov 01 04:12:44.692 INFO kablam! warning: unused imports: `RefMut`, `Ref` Nov 01 04:12:44.692 INFO kablam! --> src/gradient.rs:5:26 Nov 01 04:12:44.692 INFO kablam! | Nov 01 04:12:44.692 INFO kablam! 5 | use std::cell::{RefCell, Ref, RefMut}; Nov 01 04:12:44.692 INFO kablam! | ^^^ ^^^^^^ Nov 01 04:12:44.692 INFO kablam! | Nov 01 04:12:44.692 INFO kablam! = note: #[warn(unused_imports)] on by default Nov 01 04:12:44.692 INFO kablam! Nov 01 04:12:44.692 INFO kablam! warning: unused import: `std::cell::RefCell` Nov 01 04:12:44.692 INFO kablam! --> src/nural.rs:5:5 Nov 01 04:12:44.692 INFO kablam! | Nov 01 04:12:44.692 INFO kablam! 5 | use std::cell::RefCell; Nov 01 04:12:44.692 INFO kablam! | ^^^^^^^^^^^^^^^^^^ Nov 01 04:12:44.692 INFO kablam! Nov 01 04:12:44.868 INFO kablam! warning: unused variable: `rows` Nov 01 04:12:44.868 INFO kablam! --> src/main.rs:21:16 Nov 01 04:12:44.868 INFO kablam! | Nov 01 04:12:44.868 INFO kablam! 21 | let (size, rows, cols) = (60_000, 28, 28); Nov 01 04:12:44.868 INFO kablam! | ^^^^ help: consider using `_rows` instead Nov 01 04:12:44.868 INFO kablam! | Nov 01 04:12:44.868 INFO kablam! = note: #[warn(unused_variables)] on by default Nov 01 04:12:44.868 INFO kablam! Nov 01 04:12:44.868 INFO kablam! warning: unused variable: `cols` Nov 01 04:12:44.868 INFO kablam! --> src/main.rs:21:22 Nov 01 04:12:44.868 INFO kablam! | Nov 01 04:12:44.868 INFO kablam! 21 | let (size, rows, cols) = (60_000, 28, 28); Nov 01 04:12:44.868 INFO kablam! | ^^^^ help: consider using `_cols` instead Nov 01 04:12:44.868 INFO kablam! Nov 01 04:12:44.868 INFO kablam! warning: unused variable: `batch_size` Nov 01 04:12:44.868 INFO kablam! --> src/main.rs:40:9 Nov 01 04:12:44.868 INFO kablam! | Nov 01 04:12:44.868 INFO kablam! 40 | let batch_size = 100; Nov 01 04:12:44.868 INFO kablam! | ^^^^^^^^^^ help: consider using `_batch_size` instead Nov 01 04:12:44.868 INFO kablam! Nov 01 04:12:44.880 INFO kablam! warning: variable does not need to be mutable Nov 01 04:12:44.880 INFO kablam! --> src/main.rs:33:9 Nov 01 04:12:44.880 INFO kablam! | Nov 01 04:12:44.880 INFO kablam! 33 | let mut Two_layer_network = two_layer_net::Two_layer_network { Nov 01 04:12:44.880 INFO kablam! | ----^^^^^^^^^^^^^^^^^ Nov 01 04:12:44.880 INFO kablam! | | Nov 01 04:12:44.880 INFO kablam! | help: remove this `mut` Nov 01 04:12:44.880 INFO kablam! | Nov 01 04:12:44.880 INFO kablam! = note: #[warn(unused_mut)] on by default Nov 01 04:12:44.880 INFO kablam! Nov 01 04:12:44.921 INFO kablam! warning: function is never used: `mean_squared_error` Nov 01 04:12:44.921 INFO kablam! --> src/lossfunc.rs:11:1 Nov 01 04:12:44.921 INFO kablam! | Nov 01 04:12:44.921 INFO kablam! 11 | pub fn mean_squared_error(y: DMatrix, t: DMatrix) -> f64 { Nov 01 04:12:44.921 INFO kablam! | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Nov 01 04:12:44.921 INFO kablam! | Nov 01 04:12:44.921 INFO kablam! = note: #[warn(dead_code)] on by default Nov 01 04:12:44.921 INFO kablam! Nov 01 04:12:44.921 INFO kablam! warning: function is never used: `numerical_gradient` Nov 01 04:12:44.921 INFO kablam! --> src/gradient.rs:8:1 Nov 01 04:12:44.921 INFO kablam! | Nov 01 04:12:44.921 INFO kablam! 8 | / pub fn numerical_gradient< Nov 01 04:12:44.921 INFO kablam! 9 | | F: Fn(&DMatrix, Nov 01 04:12:44.921 INFO kablam! 10 | | &DMatrix, Nov 01 04:12:44.921 INFO kablam! 11 | | &DMatrix, Nov 01 04:12:44.921 INFO kablam! ... | Nov 01 04:12:44.921 INFO kablam! 38 | | grad Nov 01 04:12:44.921 INFO kablam! 39 | | } Nov 01 04:12:44.921 INFO kablam! | |_^ Nov 01 04:12:44.921 INFO kablam! Nov 01 04:12:44.921 INFO kablam! warning: function is never used: `function_2` Nov 01 04:12:44.921 INFO kablam! --> src/gradient.rs:40:1 Nov 01 04:12:44.921 INFO kablam! | Nov 01 04:12:44.921 INFO kablam! 40 | pub fn function_2(x: &mut DMatrix) -> f64 { Nov 01 04:12:44.921 INFO kablam! | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Nov 01 04:12:44.921 INFO kablam! Nov 01 04:12:44.921 INFO kablam! warning: method is never used: `predict` Nov 01 04:12:44.921 INFO kablam! --> src/nural.rs:11:5 Nov 01 04:12:44.921 INFO kablam! | Nov 01 04:12:44.921 INFO kablam! 11 | pub fn predict(self, x: &DMatrix) -> DMatrix { Nov 01 04:12:44.921 INFO kablam! | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Nov 01 04:12:44.922 INFO kablam! Nov 01 04:12:44.922 INFO kablam! warning: method is never used: `loss` Nov 01 04:12:44.922 INFO kablam! --> src/nural.rs:14:5 Nov 01 04:12:44.922 INFO kablam! | Nov 01 04:12:44.922 INFO kablam! 14 | pub fn loss(self, x: &DMatrix, t: &DMatrix) -> f64 { Nov 01 04:12:44.922 INFO kablam! | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Nov 01 04:12:44.922 INFO kablam! Nov 01 04:12:44.922 INFO kablam! warning: method is never used: `numerical_gradient` Nov 01 04:12:44.922 INFO kablam! --> src/two_layer_net.rs:91:5 Nov 01 04:12:44.922 INFO kablam! | Nov 01 04:12:44.922 INFO kablam! 91 | pub fn numerical_gradient(&mut self, x: &DMatrix, t: &DMatrix) -> grad { Nov 01 04:12:44.922 INFO kablam! | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Nov 01 04:12:44.922 INFO kablam! Nov 01 04:12:44.922 INFO kablam! warning: function is never used: `loss_w` Nov 01 04:12:44.922 INFO kablam! --> src/two_layer_net.rs:128:1 Nov 01 04:12:44.922 INFO kablam! | Nov 01 04:12:44.922 INFO kablam! 128 | / pub fn loss_w( Nov 01 04:12:44.922 INFO kablam! 129 | | param: &DMatrix, Nov 01 04:12:44.922 INFO kablam! 130 | | x: &DMatrix, Nov 01 04:12:44.922 INFO kablam! 131 | | t: &DMatrix, Nov 01 04:12:44.922 INFO kablam! ... | Nov 01 04:12:44.922 INFO kablam! 135 | | two.loss(param, x, t, &patern) Nov 01 04:12:44.922 INFO kablam! 136 | | } Nov 01 04:12:44.922 INFO kablam! | |_^ Nov 01 04:12:44.922 INFO kablam! Nov 01 04:12:44.922 INFO kablam! warning: function `axisZerosum` should have a snake case name such as `axis_zerosum` Nov 01 04:12:44.922 INFO kablam! --> src/nural.rs:32:1 Nov 01 04:12:44.922 INFO kablam! | Nov 01 04:12:44.922 INFO kablam! 32 | / pub fn axisZerosum(x: &DMatrix) -> DMatrix { Nov 01 04:12:44.922 INFO kablam! 33 | | let zerosum = DMatrix::::from_iterator( Nov 01 04:12:44.922 INFO kablam! 34 | | 1, Nov 01 04:12:44.922 INFO kablam! 35 | | x.shape().1, Nov 01 04:12:44.922 INFO kablam! ... | Nov 01 04:12:44.922 INFO kablam! 44 | | Nov 01 04:12:44.922 INFO kablam! 45 | | } Nov 01 04:12:44.922 INFO kablam! | |_^ Nov 01 04:12:44.922 INFO kablam! | Nov 01 04:12:44.922 INFO kablam! = note: #[warn(non_snake_case)] on by default Nov 01 04:12:44.924 INFO kablam! Nov 01 04:12:44.924 INFO kablam! warning: function `createVec` should have a snake case name such as `create_vec` Nov 01 04:12:44.924 INFO kablam! --> src/nural.rs:47:1 Nov 01 04:12:44.924 INFO kablam! | Nov 01 04:12:44.924 INFO kablam! 47 | / pub fn createVec(x: usize) -> Vec { Nov 01 04:12:44.924 INFO kablam! 48 | | let mut vec = Vec::with_capacity(x); Nov 01 04:12:44.924 INFO kablam! 49 | | for i in 0..x { Nov 01 04:12:44.924 INFO kablam! 50 | | vec.push(i); Nov 01 04:12:44.924 INFO kablam! 51 | | } Nov 01 04:12:44.924 INFO kablam! 52 | | vec Nov 01 04:12:44.924 INFO kablam! 53 | | } Nov 01 04:12:44.924 INFO kablam! | |_^ Nov 01 04:12:44.924 INFO kablam! Nov 01 04:12:44.924 INFO kablam! warning: type `grad` should have a camel case name such as `Grad` Nov 01 04:12:44.924 INFO kablam! --> src/two_layer_net.rs:13:1 Nov 01 04:12:44.924 INFO kablam! | Nov 01 04:12:44.924 INFO kablam! 13 | / pub struct grad { Nov 01 04:12:44.924 INFO kablam! 14 | | pub w1: DMatrix, Nov 01 04:12:44.924 INFO kablam! 15 | | pub b1: DMatrix, Nov 01 04:12:44.925 INFO kablam! 16 | | pub w2: DMatrix, Nov 01 04:12:44.925 INFO kablam! 17 | | pub b2: DMatrix, Nov 01 04:12:44.925 INFO kablam! 18 | | } Nov 01 04:12:44.925 INFO kablam! | |_^ Nov 01 04:12:44.925 INFO kablam! | Nov 01 04:12:44.925 INFO kablam! = note: #[warn(non_camel_case_types)] on by default Nov 01 04:12:44.925 INFO kablam! Nov 01 04:12:44.925 INFO kablam! warning: type `Two_layer_network` should have a camel case name such as `TwoLayerNetwork` Nov 01 04:12:44.925 INFO kablam! --> src/two_layer_net.rs:19:1 Nov 01 04:12:44.925 INFO kablam! | Nov 01 04:12:44.925 INFO kablam! 19 | / pub struct Two_layer_network { Nov 01 04:12:44.925 INFO kablam! 20 | | pub w1: Rc>>, Nov 01 04:12:44.925 INFO kablam! 21 | | pub b1: Rc>>, Nov 01 04:12:44.925 INFO kablam! 22 | | pub w2: Rc>>, Nov 01 04:12:44.925 INFO kablam! 23 | | pub b2: Rc>>, Nov 01 04:12:44.925 INFO kablam! 24 | | } Nov 01 04:12:44.925 INFO kablam! | |_^ Nov 01 04:12:44.925 INFO kablam! Nov 01 04:12:44.925 INFO kablam! warning: variable `Two_layer_network` should have a snake case name such as `two_layer_network` Nov 01 04:12:44.925 INFO kablam! --> src/main.rs:33:9 Nov 01 04:12:44.925 INFO kablam! | Nov 01 04:12:44.925 INFO kablam! 33 | let mut Two_layer_network = two_layer_net::Two_layer_network { Nov 01 04:12:44.925 INFO kablam! | ^^^^^^^^^^^^^^^^^^^^^ Nov 01 04:12:44.925 INFO kablam! Nov 01 04:12:44.969 INFO kablam! Finished dev [unoptimized + debuginfo] target(s) in 47.03s Nov 01 04:12:44.980 INFO kablam! su: No module specific data is present Nov 01 04:12:45.872 INFO running `"docker" "rm" "-f" "4131e103227979b3f1eba89abfc82e7cdb552d71d87f730dd5b2e604acd4db9e"` Nov 01 04:12:46.204 INFO blam! 4131e103227979b3f1eba89abfc82e7cdb552d71d87f730dd5b2e604acd4db9e