Nov 06 21:00:59.507 INFO checking helloooooo/learn_deep_learning against master#24e66c28980442a48d9458f1a4f9b76cc722dc8a for pr-55687 Nov 06 21:00:59.507 INFO running `"docker" "create" "-v" "/mnt/big/crater/work/local/target-dirs/pr-55687/worker-7/master#24e66c28980442a48d9458f1a4f9b76cc722dc8a:/target:rw,Z" "-v" "/mnt/big/crater/work/local/test-source/worker-7/pr-55687/master#24e66c28980442a48d9458f1a4f9b76cc722dc8a:/source:ro,Z" "-v" "/mnt/big/crater/work/local/cargo-home:/cargo-home:ro,Z" "-v" "/mnt/big/crater/work/local/rustup-home:/rustup-home:ro,Z" "-e" "USER_ID=1000" "-e" "SOURCE_DIR=/source" "-e" "USER_ID=1000" "-e" "CMD=cargo +24e66c28980442a48d9458f1a4f9b76cc722dc8a-alt check --frozen --all --all-targets" "-e" "CARGO_TARGET_DIR=/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/cargo-home" "-e" "RUSTUP_HOME=/rustup-home" "-m" "1536M" "--network" "none" "crater"` Nov 06 21:01:00.192 INFO blam! 8de4710fe864c02359c695b081f711e977263b2f906b4a99b73bdb9602ced7ff Nov 06 21:01:00.204 INFO running `"docker" "start" "-a" "8de4710fe864c02359c695b081f711e977263b2f906b4a99b73bdb9602ced7ff"` Nov 06 21:01:01.848 INFO kablam! usermod: no changes Nov 06 21:01:02.191 INFO kablam! Checking gnuplot v0.0.23 Nov 06 21:01:02.204 INFO kablam! Checking memchr v0.1.11 Nov 06 21:01:02.210 INFO kablam! Checking alga v0.5.2 Nov 06 21:01:02.210 INFO kablam! Checking thread-id v2.0.0 Nov 06 21:01:02.211 INFO kablam! Checking chrono v0.4.0 Nov 06 21:01:02.211 INFO kablam! Checking sha-1 v0.4.1 Nov 06 21:01:03.856 INFO kablam! Checking thread_local v0.2.7 Nov 06 21:01:04.188 INFO kablam! Checking aho-corasick v0.5.3 Nov 06 21:01:07.077 INFO kablam! Checking regex v0.1.80 Nov 06 21:01:12.080 INFO kablam! Checking egg-mode v0.12.0 Nov 06 21:01:13.112 INFO kablam! Checking nalgebra v0.13.1 Nov 06 21:01:45.697 INFO kablam! Checking test1 v0.1.0 (/source) Nov 06 21:01:46.892 INFO kablam! warning: unused imports: `RefMut`, `Ref` Nov 06 21:01:46.892 INFO kablam! --> src/gradient.rs:5:26 Nov 06 21:01:46.892 INFO kablam! | Nov 06 21:01:46.892 INFO kablam! 5 | use std::cell::{RefCell, Ref, RefMut}; Nov 06 21:01:46.892 INFO kablam! | ^^^ ^^^^^^ Nov 06 21:01:46.892 INFO kablam! | Nov 06 21:01:46.904 INFO kablam! = note: #[warn(unused_imports)] on by default Nov 06 21:01:46.904 INFO kablam! Nov 06 21:01:46.904 INFO kablam! warning: unused import: `std::cell::RefCell` Nov 06 21:01:46.904 INFO kablam! --> src/nural.rs:5:5 Nov 06 21:01:46.904 INFO kablam! | Nov 06 21:01:46.904 INFO kablam! 5 | use std::cell::RefCell; Nov 06 21:01:46.904 INFO kablam! | ^^^^^^^^^^^^^^^^^^ Nov 06 21:01:46.904 INFO kablam! Nov 06 21:01:47.013 INFO kablam! warning: unused imports: `RefMut`, `Ref` Nov 06 21:01:47.013 INFO kablam! --> src/gradient.rs:5:26 Nov 06 21:01:47.013 INFO kablam! | Nov 06 21:01:47.014 INFO kablam! 5 | use std::cell::{RefCell, Ref, RefMut}; Nov 06 21:01:47.014 INFO kablam! | ^^^ ^^^^^^ Nov 06 21:01:47.014 INFO kablam! | Nov 06 21:01:47.014 INFO kablam! = note: #[warn(unused_imports)] on by default Nov 06 21:01:47.014 INFO kablam! Nov 06 21:01:47.014 INFO kablam! warning: unused import: `std::cell::RefCell` Nov 06 21:01:47.015 INFO kablam! --> src/nural.rs:5:5 Nov 06 21:01:47.015 INFO kablam! | Nov 06 21:01:47.015 INFO kablam! 5 | use std::cell::RefCell; Nov 06 21:01:47.015 INFO kablam! | ^^^^^^^^^^^^^^^^^^ Nov 06 21:01:47.015 INFO kablam! Nov 06 21:01:47.197 INFO kablam! warning: unused variable: `rows` Nov 06 21:01:47.197 INFO kablam! --> src/main.rs:21:16 Nov 06 21:01:47.197 INFO kablam! | Nov 06 21:01:47.197 INFO kablam! 21 | let (size, rows, cols) = (60_000, 28, 28); Nov 06 21:01:47.198 INFO kablam! | ^^^^ help: consider using `_rows` instead Nov 06 21:01:47.198 INFO kablam! | Nov 06 21:01:47.198 INFO kablam! = note: #[warn(unused_variables)] on by default Nov 06 21:01:47.198 INFO kablam! Nov 06 21:01:47.198 INFO kablam! warning: unused variable: `cols` Nov 06 21:01:47.198 INFO kablam! --> src/main.rs:21:22 Nov 06 21:01:47.199 INFO kablam! | Nov 06 21:01:47.199 INFO kablam! 21 | let (size, rows, cols) = (60_000, 28, 28); Nov 06 21:01:47.199 INFO kablam! | ^^^^ help: consider using `_cols` instead Nov 06 21:01:47.199 INFO kablam! Nov 06 21:01:47.199 INFO kablam! warning: unused variable: `batch_size` Nov 06 21:01:47.200 INFO kablam! --> src/main.rs:40:9 Nov 06 21:01:47.200 INFO kablam! | Nov 06 21:01:47.200 INFO kablam! 40 | let batch_size = 100; Nov 06 21:01:47.200 INFO kablam! | ^^^^^^^^^^ help: consider using `_batch_size` instead Nov 06 21:01:47.200 INFO kablam! Nov 06 21:01:47.208 INFO kablam! warning: variable does not need to be mutable Nov 06 21:01:47.209 INFO kablam! --> src/main.rs:33:9 Nov 06 21:01:47.209 INFO kablam! | Nov 06 21:01:47.209 INFO kablam! 33 | let mut Two_layer_network = two_layer_net::Two_layer_network { Nov 06 21:01:47.210 INFO kablam! | ----^^^^^^^^^^^^^^^^^ Nov 06 21:01:47.210 INFO kablam! | | Nov 06 21:01:47.216 INFO kablam! | help: remove this `mut` Nov 06 21:01:47.216 INFO kablam! | Nov 06 21:01:47.216 INFO kablam! = note: #[warn(unused_mut)] on by default Nov 06 21:01:47.216 INFO kablam! Nov 06 21:01:47.243 INFO kablam! warning: function is never used: `mean_squared_error` Nov 06 21:01:47.244 INFO kablam! --> src/lossfunc.rs:11:1 Nov 06 21:01:47.244 INFO kablam! | Nov 06 21:01:47.244 INFO kablam! 11 | pub fn mean_squared_error(y: DMatrix, t: DMatrix) -> f64 { Nov 06 21:01:47.244 INFO kablam! | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Nov 06 21:01:47.244 INFO kablam! | Nov 06 21:01:47.244 INFO kablam! = note: #[warn(dead_code)] on by default Nov 06 21:01:47.245 INFO kablam! Nov 06 21:01:47.245 INFO kablam! warning: function is never used: `numerical_gradient` Nov 06 21:01:47.245 INFO kablam! --> src/gradient.rs:8:1 Nov 06 21:01:47.245 INFO kablam! | Nov 06 21:01:47.245 INFO kablam! 8 | / pub fn numerical_gradient< Nov 06 21:01:47.245 INFO kablam! 9 | | F: Fn(&DMatrix, Nov 06 21:01:47.245 INFO kablam! 10 | | &DMatrix, Nov 06 21:01:47.245 INFO kablam! 11 | | &DMatrix, Nov 06 21:01:47.245 INFO kablam! ... | Nov 06 21:01:47.245 INFO kablam! 38 | | grad Nov 06 21:01:47.245 INFO kablam! 39 | | } Nov 06 21:01:47.245 INFO kablam! | |_^ Nov 06 21:01:47.245 INFO kablam! Nov 06 21:01:47.245 INFO kablam! warning: function is never used: `function_2` Nov 06 21:01:47.245 INFO kablam! --> src/gradient.rs:40:1 Nov 06 21:01:47.245 INFO kablam! | Nov 06 21:01:47.245 INFO kablam! 40 | pub fn function_2(x: &mut DMatrix) -> f64 { Nov 06 21:01:47.245 INFO kablam! | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Nov 06 21:01:47.245 INFO kablam! Nov 06 21:01:47.245 INFO kablam! warning: method is never used: `predict` Nov 06 21:01:47.245 INFO kablam! --> src/nural.rs:11:5 Nov 06 21:01:47.245 INFO kablam! | Nov 06 21:01:47.245 INFO kablam! 11 | pub fn predict(self, x: &DMatrix) -> DMatrix { Nov 06 21:01:47.245 INFO kablam! | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Nov 06 21:01:47.245 INFO kablam! Nov 06 21:01:47.245 INFO kablam! warning: method is never used: `loss` Nov 06 21:01:47.245 INFO kablam! --> src/nural.rs:14:5 Nov 06 21:01:47.245 INFO kablam! | Nov 06 21:01:47.245 INFO kablam! 14 | pub fn loss(self, x: &DMatrix, t: &DMatrix) -> f64 { Nov 06 21:01:47.245 INFO kablam! | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Nov 06 21:01:47.245 INFO kablam! Nov 06 21:01:47.245 INFO kablam! warning: method is never used: `numerical_gradient` Nov 06 21:01:47.245 INFO kablam! --> src/two_layer_net.rs:91:5 Nov 06 21:01:47.245 INFO kablam! | Nov 06 21:01:47.245 INFO kablam! 91 | pub fn numerical_gradient(&mut self, x: &DMatrix, t: &DMatrix) -> grad { Nov 06 21:01:47.245 INFO kablam! | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Nov 06 21:01:47.245 INFO kablam! Nov 06 21:01:47.245 INFO kablam! warning: function is never used: `loss_w` Nov 06 21:01:47.245 INFO kablam! --> src/two_layer_net.rs:128:1 Nov 06 21:01:47.245 INFO kablam! | Nov 06 21:01:47.245 INFO kablam! 128 | / pub fn loss_w( Nov 06 21:01:47.245 INFO kablam! 129 | | param: &DMatrix, Nov 06 21:01:47.245 INFO kablam! 130 | | x: &DMatrix, Nov 06 21:01:47.245 INFO kablam! 131 | | t: &DMatrix, Nov 06 21:01:47.245 INFO kablam! ... | Nov 06 21:01:47.245 INFO kablam! 135 | | two.loss(param, x, t, &patern) Nov 06 21:01:47.245 INFO kablam! 136 | | } Nov 06 21:01:47.245 INFO kablam! | |_^ Nov 06 21:01:47.245 INFO kablam! Nov 06 21:01:47.248 INFO kablam! warning: function `axisZerosum` should have a snake case name such as `axis_zerosum` Nov 06 21:01:47.248 INFO kablam! --> src/nural.rs:32:1 Nov 06 21:01:47.248 INFO kablam! | Nov 06 21:01:47.248 INFO kablam! 32 | / pub fn axisZerosum(x: &DMatrix) -> DMatrix { Nov 06 21:01:47.248 INFO kablam! 33 | | let zerosum = DMatrix::::from_iterator( Nov 06 21:01:47.248 INFO kablam! 34 | | 1, Nov 06 21:01:47.248 INFO kablam! 35 | | x.shape().1, Nov 06 21:01:47.248 INFO kablam! ... | Nov 06 21:01:47.248 INFO kablam! 44 | | Nov 06 21:01:47.248 INFO kablam! 45 | | } Nov 06 21:01:47.248 INFO kablam! | |_^ Nov 06 21:01:47.248 INFO kablam! | Nov 06 21:01:47.248 INFO kablam! = note: #[warn(non_snake_case)] on by default Nov 06 21:01:47.248 INFO kablam! Nov 06 21:01:47.248 INFO kablam! warning: function `createVec` should have a snake case name such as `create_vec` Nov 06 21:01:47.248 INFO kablam! --> src/nural.rs:47:1 Nov 06 21:01:47.248 INFO kablam! | Nov 06 21:01:47.248 INFO kablam! 47 | / pub fn createVec(x: usize) -> Vec { Nov 06 21:01:47.249 INFO kablam! 48 | | let mut vec = Vec::with_capacity(x); Nov 06 21:01:47.249 INFO kablam! 49 | | for i in 0..x { Nov 06 21:01:47.249 INFO kablam! 50 | | vec.push(i); Nov 06 21:01:47.249 INFO kablam! 51 | | } Nov 06 21:01:47.249 INFO kablam! 52 | | vec Nov 06 21:01:47.249 INFO kablam! 53 | | } Nov 06 21:01:47.249 INFO kablam! | |_^ Nov 06 21:01:47.249 INFO kablam! Nov 06 21:01:47.249 INFO kablam! warning: type `grad` should have a camel case name such as `Grad` Nov 06 21:01:47.249 INFO kablam! --> src/two_layer_net.rs:13:1 Nov 06 21:01:47.249 INFO kablam! | Nov 06 21:01:47.249 INFO kablam! 13 | / pub struct grad { Nov 06 21:01:47.249 INFO kablam! 14 | | pub w1: DMatrix, Nov 06 21:01:47.249 INFO kablam! 15 | | pub b1: DMatrix, Nov 06 21:01:47.249 INFO kablam! 16 | | pub w2: DMatrix, Nov 06 21:01:47.249 INFO kablam! 17 | | pub b2: DMatrix, Nov 06 21:01:47.249 INFO kablam! 18 | | } Nov 06 21:01:47.249 INFO kablam! | |_^ Nov 06 21:01:47.249 INFO kablam! | Nov 06 21:01:47.249 INFO kablam! = note: #[warn(non_camel_case_types)] on by default Nov 06 21:01:47.249 INFO kablam! Nov 06 21:01:47.249 INFO kablam! warning: type `Two_layer_network` should have a camel case name such as `TwoLayerNetwork` Nov 06 21:01:47.249 INFO kablam! --> src/two_layer_net.rs:19:1 Nov 06 21:01:47.249 INFO kablam! | Nov 06 21:01:47.249 INFO kablam! 19 | / pub struct Two_layer_network { Nov 06 21:01:47.249 INFO kablam! 20 | | pub w1: Rc>>, Nov 06 21:01:47.249 INFO kablam! 21 | | pub b1: Rc>>, Nov 06 21:01:47.249 INFO kablam! 22 | | pub w2: Rc>>, Nov 06 21:01:47.249 INFO kablam! 23 | | pub b2: Rc>>, Nov 06 21:01:47.249 INFO kablam! 24 | | } Nov 06 21:01:47.249 INFO kablam! | |_^ Nov 06 21:01:47.249 INFO kablam! Nov 06 21:01:47.249 INFO kablam! warning: variable `Two_layer_network` should have a snake case name such as `two_layer_network` Nov 06 21:01:47.249 INFO kablam! --> src/main.rs:33:9 Nov 06 21:01:47.249 INFO kablam! | Nov 06 21:01:47.249 INFO kablam! 33 | let mut Two_layer_network = two_layer_net::Two_layer_network { Nov 06 21:01:47.249 INFO kablam! | ^^^^^^^^^^^^^^^^^^^^^ Nov 06 21:01:47.249 INFO kablam! Nov 06 21:01:47.277 INFO kablam! warning: unused variable: `rows` Nov 06 21:01:47.277 INFO kablam! --> src/main.rs:21:16 Nov 06 21:01:47.277 INFO kablam! | Nov 06 21:01:47.277 INFO kablam! 21 | let (size, rows, cols) = (60_000, 28, 28); Nov 06 21:01:47.277 INFO kablam! | ^^^^ help: consider using `_rows` instead Nov 06 21:01:47.277 INFO kablam! | Nov 06 21:01:47.277 INFO kablam! = note: #[warn(unused_variables)] on by default Nov 06 21:01:47.277 INFO kablam! Nov 06 21:01:47.278 INFO kablam! warning: unused variable: `cols` Nov 06 21:01:47.278 INFO kablam! --> src/main.rs:21:22 Nov 06 21:01:47.278 INFO kablam! | Nov 06 21:01:47.278 INFO kablam! 21 | let (size, rows, cols) = (60_000, 28, 28); Nov 06 21:01:47.278 INFO kablam! | ^^^^ help: consider using `_cols` instead Nov 06 21:01:47.278 INFO kablam! Nov 06 21:01:47.278 INFO kablam! warning: unused variable: `batch_size` Nov 06 21:01:47.278 INFO kablam! --> src/main.rs:40:9 Nov 06 21:01:47.278 INFO kablam! | Nov 06 21:01:47.278 INFO kablam! 40 | let batch_size = 100; Nov 06 21:01:47.278 INFO kablam! | ^^^^^^^^^^ help: consider using `_batch_size` instead Nov 06 21:01:47.281 INFO kablam! Nov 06 21:01:47.295 INFO kablam! warning: variable does not need to be mutable Nov 06 21:01:47.295 INFO kablam! --> src/main.rs:33:9 Nov 06 21:01:47.295 INFO kablam! | Nov 06 21:01:47.295 INFO kablam! 33 | let mut Two_layer_network = two_layer_net::Two_layer_network { Nov 06 21:01:47.295 INFO kablam! | ----^^^^^^^^^^^^^^^^^ Nov 06 21:01:47.295 INFO kablam! | | Nov 06 21:01:47.295 INFO kablam! | help: remove this `mut` Nov 06 21:01:47.296 INFO kablam! | Nov 06 21:01:47.296 INFO kablam! = note: #[warn(unused_mut)] on by default Nov 06 21:01:47.296 INFO kablam! Nov 06 21:01:47.369 INFO kablam! warning: function is never used: `numerical_gradient` Nov 06 21:01:47.369 INFO kablam! --> src/gradient.rs:8:1 Nov 06 21:01:47.369 INFO kablam! | Nov 06 21:01:47.369 INFO kablam! 8 | / pub fn numerical_gradient< Nov 06 21:01:47.369 INFO kablam! 9 | | F: Fn(&DMatrix, Nov 06 21:01:47.369 INFO kablam! 10 | | &DMatrix, Nov 06 21:01:47.369 INFO kablam! 11 | | &DMatrix, Nov 06 21:01:47.369 INFO kablam! ... | Nov 06 21:01:47.369 INFO kablam! 38 | | grad Nov 06 21:01:47.369 INFO kablam! 39 | | } Nov 06 21:01:47.369 INFO kablam! | |_^ Nov 06 21:01:47.369 INFO kablam! | Nov 06 21:01:47.370 INFO kablam! = note: #[warn(dead_code)] on by default Nov 06 21:01:47.370 INFO kablam! Nov 06 21:01:47.370 INFO kablam! warning: method is never used: `numerical_gradient` Nov 06 21:01:47.370 INFO kablam! --> src/two_layer_net.rs:91:5 Nov 06 21:01:47.370 INFO kablam! | Nov 06 21:01:47.370 INFO kablam! 91 | pub fn numerical_gradient(&mut self, x: &DMatrix, t: &DMatrix) -> grad { Nov 06 21:01:47.370 INFO kablam! | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Nov 06 21:01:47.370 INFO kablam! Nov 06 21:01:47.370 INFO kablam! warning: function is never used: `loss_w` Nov 06 21:01:47.370 INFO kablam! --> src/two_layer_net.rs:128:1 Nov 06 21:01:47.371 INFO kablam! | Nov 06 21:01:47.371 INFO kablam! 128 | / pub fn loss_w( Nov 06 21:01:47.371 INFO kablam! 129 | | param: &DMatrix, Nov 06 21:01:47.373 INFO kablam! 130 | | x: &DMatrix, Nov 06 21:01:47.373 INFO kablam! 131 | | t: &DMatrix, Nov 06 21:01:47.373 INFO kablam! ... | Nov 06 21:01:47.373 INFO kablam! 135 | | two.loss(param, x, t, &patern) Nov 06 21:01:47.373 INFO kablam! 136 | | } Nov 06 21:01:47.373 INFO kablam! | |_^ Nov 06 21:01:47.373 INFO kablam! Nov 06 21:01:47.373 INFO kablam! warning: function `axisZerosum` should have a snake case name such as `axis_zerosum` Nov 06 21:01:47.373 INFO kablam! --> src/nural.rs:32:1 Nov 06 21:01:47.373 INFO kablam! | Nov 06 21:01:47.373 INFO kablam! 32 | / pub fn axisZerosum(x: &DMatrix) -> DMatrix { Nov 06 21:01:47.373 INFO kablam! 33 | | let zerosum = DMatrix::::from_iterator( Nov 06 21:01:47.373 INFO kablam! 34 | | 1, Nov 06 21:01:47.373 INFO kablam! 35 | | x.shape().1, Nov 06 21:01:47.373 INFO kablam! ... | Nov 06 21:01:47.373 INFO kablam! 44 | | Nov 06 21:01:47.373 INFO kablam! 45 | | } Nov 06 21:01:47.373 INFO kablam! | |_^ Nov 06 21:01:47.373 INFO kablam! | Nov 06 21:01:47.373 INFO kablam! = note: #[warn(non_snake_case)] on by default Nov 06 21:01:47.373 INFO kablam! Nov 06 21:01:47.373 INFO kablam! warning: function `createVec` should have a snake case name such as `create_vec` Nov 06 21:01:47.373 INFO kablam! --> src/nural.rs:47:1 Nov 06 21:01:47.373 INFO kablam! | Nov 06 21:01:47.373 INFO kablam! 47 | / pub fn createVec(x: usize) -> Vec { Nov 06 21:01:47.373 INFO kablam! 48 | | let mut vec = Vec::with_capacity(x); Nov 06 21:01:47.373 INFO kablam! 49 | | for i in 0..x { Nov 06 21:01:47.373 INFO kablam! 50 | | vec.push(i); Nov 06 21:01:47.374 INFO kablam! 51 | | } Nov 06 21:01:47.374 INFO kablam! 52 | | vec Nov 06 21:01:47.374 INFO kablam! 53 | | } Nov 06 21:01:47.374 INFO kablam! | |_^ Nov 06 21:01:47.374 INFO kablam! Nov 06 21:01:47.374 INFO kablam! warning: type `grad` should have a camel case name such as `Grad` Nov 06 21:01:47.374 INFO kablam! --> src/two_layer_net.rs:13:1 Nov 06 21:01:47.374 INFO kablam! | Nov 06 21:01:47.374 INFO kablam! 13 | / pub struct grad { Nov 06 21:01:47.374 INFO kablam! 14 | | pub w1: DMatrix, Nov 06 21:01:47.374 INFO kablam! 15 | | pub b1: DMatrix, Nov 06 21:01:47.374 INFO kablam! 16 | | pub w2: DMatrix, Nov 06 21:01:47.374 INFO kablam! 17 | | pub b2: DMatrix, Nov 06 21:01:47.374 INFO kablam! 18 | | } Nov 06 21:01:47.374 INFO kablam! | |_^ Nov 06 21:01:47.374 INFO kablam! | Nov 06 21:01:47.374 INFO kablam! = note: #[warn(non_camel_case_types)] on by default Nov 06 21:01:47.374 INFO kablam! Nov 06 21:01:47.374 INFO kablam! warning: type `Two_layer_network` should have a camel case name such as `TwoLayerNetwork` Nov 06 21:01:47.374 INFO kablam! --> src/two_layer_net.rs:19:1 Nov 06 21:01:47.374 INFO kablam! | Nov 06 21:01:47.374 INFO kablam! 19 | / pub struct Two_layer_network { Nov 06 21:01:47.374 INFO kablam! 20 | | pub w1: Rc>>, Nov 06 21:01:47.374 INFO kablam! 21 | | pub b1: Rc>>, Nov 06 21:01:47.374 INFO kablam! 22 | | pub w2: Rc>>, Nov 06 21:01:47.374 INFO kablam! 23 | | pub b2: Rc>>, Nov 06 21:01:47.374 INFO kablam! 24 | | } Nov 06 21:01:47.375 INFO kablam! | |_^ Nov 06 21:01:47.375 INFO kablam! Nov 06 21:01:47.375 INFO kablam! warning: variable `Two_layer_network` should have a snake case name such as `two_layer_network` Nov 06 21:01:47.375 INFO kablam! --> src/main.rs:33:9 Nov 06 21:01:47.375 INFO kablam! | Nov 06 21:01:47.375 INFO kablam! 33 | let mut Two_layer_network = two_layer_net::Two_layer_network { Nov 06 21:01:47.375 INFO kablam! | ^^^^^^^^^^^^^^^^^^^^^ Nov 06 21:01:47.375 INFO kablam! Nov 06 21:01:47.425 INFO kablam! Finished dev [unoptimized + debuginfo] target(s) in 45.55s Nov 06 21:01:47.456 INFO kablam! su: No module specific data is present Nov 06 21:01:48.254 INFO running `"docker" "rm" "-f" "8de4710fe864c02359c695b081f711e977263b2f906b4a99b73bdb9602ced7ff"` Nov 06 21:01:48.612 INFO blam! 8de4710fe864c02359c695b081f711e977263b2f906b4a99b73bdb9602ced7ff