[INFO] fetching crate collenchyma-nn 0.3.4...
[INFO] checking collenchyma-nn-0.3.4 against master#30f74ff0dc4d66debc8b50724c446f817e5f75f4 for pr-121848-1
[INFO] extracting crate collenchyma-nn 0.3.4 into /workspace/builds/worker-2-tc1/source
[INFO] validating manifest of crates.io crate collenchyma-nn 0.3.4 on toolchain 30f74ff0dc4d66debc8b50724c446f817e5f75f4
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+30f74ff0dc4d66debc8b50724c446f817e5f75f4" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }`
[INFO] started tweaking crates.io crate collenchyma-nn 0.3.4
[INFO] finished tweaking crates.io crate collenchyma-nn 0.3.4
[INFO] tweaked toml for crates.io crate collenchyma-nn 0.3.4 written to /workspace/builds/worker-2-tc1/source/Cargo.toml
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+30f74ff0dc4d66debc8b50724c446f817e5f75f4" "generate-lockfile" "--manifest-path" "Cargo.toml", kill_on_drop: false }`
[INFO] [stderr]     Blocking waiting for file lock on package cache
[INFO] [stderr]     Updating crates.io index
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+30f74ff0dc4d66debc8b50724c446f817e5f75f4" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }`
[INFO] [stderr]     Blocking waiting for file lock on package cache
[INFO] [stderr]  Downloading crates ...
[INFO] [stderr]   Downloaded linear-map v0.0.4
[INFO] [stderr]   Downloaded bitflags v0.3.3
[INFO] [stderr]   Downloaded byteorder v0.4.2
[INFO] [stderr]   Downloaded collenchyma v0.0.8
[INFO] [stderr]   Downloaded clippy v0.0.27
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-2-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-2-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:2788e3201cd34a07e3172128adcd8b3090168a8e3bcc40d7c032b9dda1df7d1c" "/opt/rustwide/cargo-home/bin/cargo" "+30f74ff0dc4d66debc8b50724c446f817e5f75f4" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }`
[INFO] [stdout] 82a6dc870efca5a94a035030214e01f62282c68c7bd40388986a0f5a924b7dab
[INFO] running `Command { std: "docker" "start" "-a" "82a6dc870efca5a94a035030214e01f62282c68c7bd40388986a0f5a924b7dab", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "inspect" "82a6dc870efca5a94a035030214e01f62282c68c7bd40388986a0f5a924b7dab", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "82a6dc870efca5a94a035030214e01f62282c68c7bd40388986a0f5a924b7dab", kill_on_drop: false }`
[INFO] [stdout] 82a6dc870efca5a94a035030214e01f62282c68c7bd40388986a0f5a924b7dab
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-2-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-2-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:2788e3201cd34a07e3172128adcd8b3090168a8e3bcc40d7c032b9dda1df7d1c" "/opt/rustwide/cargo-home/bin/cargo" "+30f74ff0dc4d66debc8b50724c446f817e5f75f4" "check" "--frozen" "--all" "--all-targets" "--message-format=json", kill_on_drop: false }`
[INFO] [stdout] 4602928a37137b580d96624737d482c60a6c5d5d62df3bf767580e1a841f2ef8
[INFO] running `Command { std: "docker" "start" "-a" "4602928a37137b580d96624737d482c60a6c5d5d62df3bf767580e1a841f2ef8", kill_on_drop: false }`
[INFO] [stderr]     Checking bitflags v0.3.3
[INFO] [stderr]     Checking byteorder v0.4.2
[INFO] [stderr]     Checking lazy_static v0.1.16
[INFO] [stderr]     Checking linear-map v0.0.4
[INFO] [stderr]     Checking collenchyma v0.0.8
[INFO] [stderr]     Checking collenchyma-nn v0.3.4 (/opt/rustwide/workdir)
[INFO] [stdout] error[E0428]: the name `impl_ops_sigmoid_for` is defined multiple times
[INFO] [stdout]   --> src/frameworks/cuda/helper.rs:47:1
[INFO] [stdout]    |
[INFO] [stdout] 47 | macro_rules! impl_ops_sigmoid_for {
[INFO] [stdout]    | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ `impl_ops_sigmoid_for` redefined here
[INFO] [stdout]    |
[INFO] [stdout]   ::: src/frameworks/native/helper.rs:92:1
[INFO] [stdout]    |
[INFO] [stdout] 92 | macro_rules! impl_ops_sigmoid_for {
[INFO] [stdout]    | --------------------------------- previous definition of the macro `impl_ops_sigmoid_for` here
[INFO] [stdout]    |
[INFO] [stdout]    = note: `impl_ops_sigmoid_for` must be defined only once in the macro namespace of this module
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0428]: the name `impl_ops_sigmoid_for` is defined multiple times
[INFO] [stdout]   --> src/frameworks/cuda/helper.rs:47:1
[INFO] [stdout]    |
[INFO] [stdout] 47 | macro_rules! impl_ops_sigmoid_for {
[INFO] [stdout]    | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ `impl_ops_sigmoid_for` redefined here
[INFO] [stdout]    |
[INFO] [stdout]   ::: src/frameworks/native/helper.rs:92:1
[INFO] [stdout]    |
[INFO] [stdout] 92 | macro_rules! impl_ops_sigmoid_for {
[INFO] [stdout]    | --------------------------------- previous definition of the macro `impl_ops_sigmoid_for` here
[INFO] [stdout]    |
[INFO] [stdout]    = note: `impl_ops_sigmoid_for` must be defined only once in the macro namespace of this module
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0428]: the name `impl_ops_relu_for` is defined multiple times
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:200:1
[INFO] [stdout]     |
[INFO] [stdout] 200 | macro_rules! impl_ops_relu_for {
[INFO] [stdout]     | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ `impl_ops_relu_for` redefined here
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/helper.rs:155:1
[INFO] [stdout]     |
[INFO] [stdout] 155 | macro_rules! impl_ops_relu_for {
[INFO] [stdout]     | ------------------------------ previous definition of the macro `impl_ops_relu_for` here
[INFO] [stdout]     |
[INFO] [stdout]     = note: `impl_ops_relu_for` must be defined only once in the macro namespace of this module
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0428]: the name `impl_ops_relu_for` is defined multiple times
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:200:1
[INFO] [stdout]     |
[INFO] [stdout] 200 | macro_rules! impl_ops_relu_for {
[INFO] [stdout]     | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ `impl_ops_relu_for` redefined here
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/helper.rs:155:1
[INFO] [stdout]     |
[INFO] [stdout] 155 | macro_rules! impl_ops_relu_for {
[INFO] [stdout]     | ------------------------------ previous definition of the macro `impl_ops_relu_for` here
[INFO] [stdout]     |
[INFO] [stdout]     = note: `impl_ops_relu_for` must be defined only once in the macro namespace of this module
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0428]: the name `impl_ops_tanh_for` is defined multiple times
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:353:1
[INFO] [stdout]     |
[INFO] [stdout] 353 | macro_rules! impl_ops_tanh_for {
[INFO] [stdout]     | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ `impl_ops_tanh_for` redefined here
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/helper.rs:218:1
[INFO] [stdout]     |
[INFO] [stdout] 218 | macro_rules! impl_ops_tanh_for {
[INFO] [stdout]     | ------------------------------ previous definition of the macro `impl_ops_tanh_for` here
[INFO] [stdout]     |
[INFO] [stdout]     = note: `impl_ops_tanh_for` must be defined only once in the macro namespace of this module
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0428]: the name `impl_ops_tanh_for` is defined multiple times
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:353:1
[INFO] [stdout]     |
[INFO] [stdout] 353 | macro_rules! impl_ops_tanh_for {
[INFO] [stdout]     | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ `impl_ops_tanh_for` redefined here
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/helper.rs:218:1
[INFO] [stdout]     |
[INFO] [stdout] 218 | macro_rules! impl_ops_tanh_for {
[INFO] [stdout]     | ------------------------------ previous definition of the macro `impl_ops_tanh_for` here
[INFO] [stdout]     |
[INFO] [stdout]     = note: `impl_ops_tanh_for` must be defined only once in the macro namespace of this module
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0428]: the name `impl_ops_convolution_for` is defined multiple times
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:506:1
[INFO] [stdout]     |
[INFO] [stdout] 506 | macro_rules! impl_ops_convolution_for {
[INFO] [stdout]     | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ `impl_ops_convolution_for` redefined here
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/helper.rs:282:1
[INFO] [stdout]     |
[INFO] [stdout] 282 | macro_rules! impl_ops_convolution_for {
[INFO] [stdout]     | ------------------------------------- previous definition of the macro `impl_ops_convolution_for` here
[INFO] [stdout]     |
[INFO] [stdout]     = note: `impl_ops_convolution_for` must be defined only once in the macro namespace of this module
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0428]: the name `impl_ops_convolution_for` is defined multiple times
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:506:1
[INFO] [stdout]     |
[INFO] [stdout] 506 | macro_rules! impl_ops_convolution_for {
[INFO] [stdout]     | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ `impl_ops_convolution_for` redefined here
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/helper.rs:282:1
[INFO] [stdout]     |
[INFO] [stdout] 282 | macro_rules! impl_ops_convolution_for {
[INFO] [stdout]     | ------------------------------------- previous definition of the macro `impl_ops_convolution_for` here
[INFO] [stdout]     |
[INFO] [stdout]     = note: `impl_ops_convolution_for` must be defined only once in the macro namespace of this module
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0428]: the name `impl_ops_softmax_for` is defined multiple times
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:643:1
[INFO] [stdout]     |
[INFO] [stdout] 643 | macro_rules! impl_ops_softmax_for {
[INFO] [stdout]     | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ `impl_ops_softmax_for` redefined here
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/helper.rs:344:1
[INFO] [stdout]     |
[INFO] [stdout] 344 | macro_rules! impl_ops_softmax_for {
[INFO] [stdout]     | --------------------------------- previous definition of the macro `impl_ops_softmax_for` here
[INFO] [stdout]     |
[INFO] [stdout]     = note: `impl_ops_softmax_for` must be defined only once in the macro namespace of this module
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0428]: the name `impl_ops_log_softmax_for` is defined multiple times
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:719:1
[INFO] [stdout]     |
[INFO] [stdout] 719 | macro_rules! impl_ops_log_softmax_for {
[INFO] [stdout]     | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ `impl_ops_log_softmax_for` redefined here
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/helper.rs:416:1
[INFO] [stdout]     |
[INFO] [stdout] 416 | macro_rules! impl_ops_log_softmax_for {
[INFO] [stdout]     | ------------------------------------- previous definition of the macro `impl_ops_log_softmax_for` here
[INFO] [stdout]     |
[INFO] [stdout]     = note: `impl_ops_log_softmax_for` must be defined only once in the macro namespace of this module
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0428]: the name `impl_ops_softmax_for` is defined multiple times
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:643:1
[INFO] [stdout]     |
[INFO] [stdout] 643 | macro_rules! impl_ops_softmax_for {
[INFO] [stdout]     | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ `impl_ops_softmax_for` redefined here
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/helper.rs:344:1
[INFO] [stdout]     |
[INFO] [stdout] 344 | macro_rules! impl_ops_softmax_for {
[INFO] [stdout]     | --------------------------------- previous definition of the macro `impl_ops_softmax_for` here
[INFO] [stdout]     |
[INFO] [stdout]     = note: `impl_ops_softmax_for` must be defined only once in the macro namespace of this module
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0428]: the name `impl_ops_lrn_for` is defined multiple times
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:795:1
[INFO] [stdout]     |
[INFO] [stdout] 795 | macro_rules! impl_ops_lrn_for {
[INFO] [stdout]     | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ `impl_ops_lrn_for` redefined here
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/helper.rs:495:1
[INFO] [stdout]     |
[INFO] [stdout] 495 | macro_rules! impl_ops_lrn_for {
[INFO] [stdout]     | ----------------------------- previous definition of the macro `impl_ops_lrn_for` here
[INFO] [stdout]     |
[INFO] [stdout]     = note: `impl_ops_lrn_for` must be defined only once in the macro namespace of this module
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0428]: the name `impl_ops_log_softmax_for` is defined multiple times
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:719:1
[INFO] [stdout]     |
[INFO] [stdout] 719 | macro_rules! impl_ops_log_softmax_for {
[INFO] [stdout]     | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ `impl_ops_log_softmax_for` redefined here
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/helper.rs:416:1
[INFO] [stdout]     |
[INFO] [stdout] 416 | macro_rules! impl_ops_log_softmax_for {
[INFO] [stdout]     | ------------------------------------- previous definition of the macro `impl_ops_log_softmax_for` here
[INFO] [stdout]     |
[INFO] [stdout]     = note: `impl_ops_log_softmax_for` must be defined only once in the macro namespace of this module
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0428]: the name `impl_ops_pooling_for` is defined multiple times
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:894:1
[INFO] [stdout]     |
[INFO] [stdout] 894 | macro_rules! impl_ops_pooling_for {
[INFO] [stdout]     | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ `impl_ops_pooling_for` redefined here
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/helper.rs:557:1
[INFO] [stdout]     |
[INFO] [stdout] 557 | macro_rules! impl_ops_pooling_for {
[INFO] [stdout]     | --------------------------------- previous definition of the macro `impl_ops_pooling_for` here
[INFO] [stdout]     |
[INFO] [stdout]     = note: `impl_ops_pooling_for` must be defined only once in the macro namespace of this module
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0428]: the name `impl_ops_lrn_for` is defined multiple times
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:795:1
[INFO] [stdout]     |
[INFO] [stdout] 795 | macro_rules! impl_ops_lrn_for {
[INFO] [stdout]     | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ `impl_ops_lrn_for` redefined here
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/helper.rs:495:1
[INFO] [stdout]     |
[INFO] [stdout] 495 | macro_rules! impl_ops_lrn_for {
[INFO] [stdout]     | ----------------------------- previous definition of the macro `impl_ops_lrn_for` here
[INFO] [stdout]     |
[INFO] [stdout]     = note: `impl_ops_lrn_for` must be defined only once in the macro namespace of this module
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0428]: the name `impl_ops_pooling_for` is defined multiple times
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:894:1
[INFO] [stdout]     |
[INFO] [stdout] 894 | macro_rules! impl_ops_pooling_for {
[INFO] [stdout]     | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ `impl_ops_pooling_for` redefined here
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/helper.rs:557:1
[INFO] [stdout]     |
[INFO] [stdout] 557 | macro_rules! impl_ops_pooling_for {
[INFO] [stdout]     | --------------------------------- previous definition of the macro `impl_ops_pooling_for` here
[INFO] [stdout]     |
[INFO] [stdout]     = note: `impl_ops_pooling_for` must be defined only once in the macro namespace of this module
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused `#[macro_use]` import
[INFO] [stdout]    --> src/lib.rs:119:1
[INFO] [stdout]     |
[INFO] [stdout] 119 | #[macro_use]
[INFO] [stdout]     | ^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout]     = note: `#[warn(unused_imports)]` on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:165:92
[INFO] [stdout]     |
[INFO] [stdout] 165 |     fn sigmoid(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                            ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] note: the lint level is defined here
[INFO] [stdout]    --> src/lib.rs:111:31
[INFO] [stdout]     |
[INFO] [stdout] 111 |         unused_import_braces, unused_qualifications)]
[INFO] [stdout]     |                               ^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 165 -     fn sigmoid(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 165 +     fn sigmoid(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:174:94
[INFO] [stdout]     |
[INFO] [stdout] 174 |     fn sigmoid_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 174 -     fn sigmoid_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 174 +     fn sigmoid_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:182:162
[INFO] [stdout]     |
[INFO] [stdout] 182 | ...nsor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 182 -     fn sigmoid_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 182 +     fn sigmoid_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:191:156
[INFO] [stdout]     |
[INFO] [stdout] 191 | ...nsor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 191 -     fn sigmoid_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result: &SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 191 +     fn sigmoid_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result: &SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:202:72
[INFO] [stdout]     |
[INFO] [stdout] 202 |     fn sigmoid_pointwise(&self, x: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                        ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 202 -     fn sigmoid_pointwise(&self, x: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 202 +     fn sigmoid_pointwise(&self, x: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:211:78
[INFO] [stdout]     |
[INFO] [stdout] 211 |     fn sigmoid_pointwise_plain(&self, x: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 211 -     fn sigmoid_pointwise_plain(&self, x: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 211 +     fn sigmoid_pointwise_plain(&self, x: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:219:107
[INFO] [stdout]     |
[INFO] [stdout] 219 |     fn sigmoid_pointwise_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                                           ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 219 -     fn sigmoid_pointwise_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 219 +     fn sigmoid_pointwise_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:228:109
[INFO] [stdout]     |
[INFO] [stdout] 228 |     fn sigmoid_pointwise_grad_plain(&self, x: &SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                                             ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 228 -     fn sigmoid_pointwise_grad_plain(&self, x: &SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 228 +     fn sigmoid_pointwise_grad_plain(&self, x: &SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:239:89
[INFO] [stdout]     |
[INFO] [stdout] 239 |     fn relu(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                         ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 239 -     fn relu(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 239 +     fn relu(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:248:91
[INFO] [stdout]     |
[INFO] [stdout] 248 |     fn relu_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                           ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 248 -     fn relu_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 248 +     fn relu_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:256:159
[INFO] [stdout]     |
[INFO] [stdout] 256 | ...nsor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 256 -     fn relu_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 256 +     fn relu_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:265:153
[INFO] [stdout]     |
[INFO] [stdout] 265 | ...nsor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 265 -     fn relu_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result: &SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 265 +     fn relu_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result: &SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:276:69
[INFO] [stdout]     |
[INFO] [stdout] 276 |     fn relu_pointwise(&self, x: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                     ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 276 -     fn relu_pointwise(&self, x: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 276 +     fn relu_pointwise(&self, x: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:285:75
[INFO] [stdout]     |
[INFO] [stdout] 285 |     fn relu_pointwise_plain(&self, x: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                           ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 285 -     fn relu_pointwise_plain(&self, x: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 285 +     fn relu_pointwise_plain(&self, x: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:293:104
[INFO] [stdout]     |
[INFO] [stdout] 293 |     fn relu_pointwise_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                                        ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 293 -     fn relu_pointwise_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 293 +     fn relu_pointwise_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:302:106
[INFO] [stdout]     |
[INFO] [stdout] 302 |     fn relu_pointwise_grad_plain(&self, x: &SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                                          ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 302 -     fn relu_pointwise_grad_plain(&self, x: &SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 302 +     fn relu_pointwise_grad_plain(&self, x: &SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:313:89
[INFO] [stdout]     |
[INFO] [stdout] 313 |     fn tanh(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                         ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 313 -     fn tanh(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 313 +     fn tanh(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:322:91
[INFO] [stdout]     |
[INFO] [stdout] 322 |     fn tanh_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                           ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 322 -     fn tanh_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 322 +     fn tanh_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:330:159
[INFO] [stdout]     |
[INFO] [stdout] 330 | ...nsor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 330 -     fn tanh_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 330 +     fn tanh_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:339:153
[INFO] [stdout]     |
[INFO] [stdout] 339 | ...nsor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 339 -     fn tanh_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result: &SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 339 +     fn tanh_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result: &SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:350:69
[INFO] [stdout]     |
[INFO] [stdout] 350 |     fn tanh_pointwise(&self, x: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                     ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 350 -     fn tanh_pointwise(&self, x: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 350 +     fn tanh_pointwise(&self, x: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:359:75
[INFO] [stdout]     |
[INFO] [stdout] 359 |     fn tanh_pointwise_plain(&self, x: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                           ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 359 -     fn tanh_pointwise_plain(&self, x: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 359 +     fn tanh_pointwise_plain(&self, x: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:367:104
[INFO] [stdout]     |
[INFO] [stdout] 367 |     fn tanh_pointwise_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                                        ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 367 -     fn tanh_pointwise_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 367 +     fn tanh_pointwise_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:376:106
[INFO] [stdout]     |
[INFO] [stdout] 376 |     fn tanh_pointwise_grad_plain(&self, x: &SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                                          ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 376 -     fn tanh_pointwise_grad_plain(&self, x: &SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 376 +     fn tanh_pointwise_grad_plain(&self, x: &SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:384:87
[INFO] [stdout]     |
[INFO] [stdout] 384 | ...                   stride: &[i32], zero_padding: &[i32]) -> Result<Self::CC, ::co::error::Error>;
[INFO] [stdout]     |                                                                                 ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 384 -                             stride: &[i32], zero_padding: &[i32]) -> Result<Self::CC, ::co::error::Error>;
[INFO] [stdout] 384 +                             stride: &[i32], zero_padding: &[i32]) -> Result<Self::CC, co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:392:179
[INFO] [stdout]     |
[INFO] [stdout] 392 | ...: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 392 -     fn convolution(&self, filter: &mut SharedTensor<F>, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>, workspace: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 392 +     fn convolution(&self, filter: &mut SharedTensor<F>, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>, workspace: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:401:177
[INFO] [stdout]     |
[INFO] [stdout] 401 | ...: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 401 -     fn convolution_plain(&self, filter: &SharedTensor<F>, x: &SharedTensor<F>, result: &mut SharedTensor<F>, workspace: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 401 +     fn convolution_plain(&self, filter: &SharedTensor<F>, x: &SharedTensor<F>, result: &mut SharedTensor<F>, workspace: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:409:206
[INFO] [stdout]     |
[INFO] [stdout] 409 | ...: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 409 -     fn convolution_grad_filter(&self, src_data: &mut SharedTensor<F>, dest_diff: &mut SharedTensor<F>, filter_diff: &mut SharedTensor<F>, workspace: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 409 +     fn convolution_grad_filter(&self, src_data: &mut SharedTensor<F>, dest_diff: &mut SharedTensor<F>, filter_diff: &mut SharedTensor<F>, workspace: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:418:204
[INFO] [stdout]     |
[INFO] [stdout] 418 | ...: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 418 -     fn convolution_grad_filter_plain(&self, src_data: &SharedTensor<F>, dest_diff: &SharedTensor<F>, filter_diff: &mut SharedTensor<F>, workspace: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 418 +     fn convolution_grad_filter_plain(&self, src_data: &SharedTensor<F>, dest_diff: &SharedTensor<F>, filter_diff: &mut SharedTensor<F>, workspace: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:426:199
[INFO] [stdout]     |
[INFO] [stdout] 426 | ...: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 426 -     fn convolution_grad_data(&self, filter: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>, workspace: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 426 +     fn convolution_grad_data(&self, filter: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>, workspace: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:435:197
[INFO] [stdout]     |
[INFO] [stdout] 435 | ...: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 435 -     fn convolution_grad_data_plain(&self, filter: &SharedTensor<F>, x_diff: &SharedTensor<F>, result_diff: &mut SharedTensor<F>, workspace: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 435 +     fn convolution_grad_data_plain(&self, filter: &SharedTensor<F>, x_diff: &SharedTensor<F>, result_diff: &mut SharedTensor<F>, workspace: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:472:92
[INFO] [stdout]     |
[INFO] [stdout] 472 |     fn softmax(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                            ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 472 -     fn softmax(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 472 +     fn softmax(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:481:94
[INFO] [stdout]     |
[INFO] [stdout] 481 |     fn softmax_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 481 -     fn softmax_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 481 +     fn softmax_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused `#[macro_use]` import
[INFO] [stdout]    --> src/lib.rs:119:1
[INFO] [stdout]     |
[INFO] [stdout] 119 | #[macro_use]
[INFO] [stdout]     | ^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout]     = note: `#[warn(unused_imports)]` on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:489:132
[INFO] [stdout]     |
[INFO] [stdout] 489 | ...nsor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 489 -     fn softmax_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 489 +     fn softmax_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:498:130
[INFO] [stdout]     |
[INFO] [stdout] 498 | ...nsor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 498 -     fn softmax_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 498 +     fn softmax_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:508:96
[INFO] [stdout]     |
[INFO] [stdout] 508 |     fn log_softmax(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                                ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 508 -     fn log_softmax(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 508 +     fn log_softmax(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:517:98
[INFO] [stdout]     |
[INFO] [stdout] 517 |     fn log_softmax_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                                  ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 517 -     fn log_softmax_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 517 +     fn log_softmax_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:165:92
[INFO] [stdout]     |
[INFO] [stdout] 165 |     fn sigmoid(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                            ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] note: the lint level is defined here
[INFO] [stdout]    --> src/lib.rs:111:31
[INFO] [stdout]     |
[INFO] [stdout] 111 |         unused_import_braces, unused_qualifications)]
[INFO] [stdout]     |                               ^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 165 -     fn sigmoid(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 165 +     fn sigmoid(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:524:136
[INFO] [stdout]     |
[INFO] [stdout] 524 | ...nsor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 524 -     fn log_softmax_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 524 +     fn log_softmax_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:174:94
[INFO] [stdout]     |
[INFO] [stdout] 174 |     fn sigmoid_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 174 -     fn sigmoid_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 174 +     fn sigmoid_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:533:134
[INFO] [stdout]     |
[INFO] [stdout] 533 | ...nsor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 533 -     fn log_softmax_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 533 +     fn log_softmax_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:182:162
[INFO] [stdout]     |
[INFO] [stdout] 182 | ...nsor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 182 -     fn sigmoid_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 182 +     fn sigmoid_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:539:91
[INFO] [stdout]     |
[INFO] [stdout] 539 |     fn new_lrn_config(&self, n: u32, alpha: f64, beta: f64, k: f64) -> Result<Self::CLRN, ::co::error::Error>;
[INFO] [stdout]     |                                                                                           ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 539 -     fn new_lrn_config(&self, n: u32, alpha: f64, beta: f64, k: f64) -> Result<Self::CLRN, ::co::error::Error>;
[INFO] [stdout] 539 +     fn new_lrn_config(&self, n: u32, alpha: f64, beta: f64, k: f64) -> Result<Self::CLRN, co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:191:156
[INFO] [stdout]     |
[INFO] [stdout] 191 | ...nsor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 191 -     fn sigmoid_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result: &SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 191 +     fn sigmoid_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result: &SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:547:109
[INFO] [stdout]     |
[INFO] [stdout] 547 |     fn lrn(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>, config: &Self::CLRN) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                                             ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 547 -     fn lrn(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>, config: &Self::CLRN) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 547 +     fn lrn(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>, config: &Self::CLRN) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:556:111
[INFO] [stdout]     |
[INFO] [stdout] 556 |     fn lrn_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>, config: &Self::CLRN) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                                               ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 556 -     fn lrn_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>, config: &Self::CLRN) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 556 +     fn lrn_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>, config: &Self::CLRN) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:202:72
[INFO] [stdout]     |
[INFO] [stdout] 202 |     fn sigmoid_pointwise(&self, x: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                        ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 202 -     fn sigmoid_pointwise(&self, x: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 202 +     fn sigmoid_pointwise(&self, x: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:564:179
[INFO] [stdout]     |
[INFO] [stdout] 564 | ... &mut SharedTensor<F>, config: &Self::CLRN) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 564 -     fn lrn_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>, config: &Self::CLRN) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 564 +     fn lrn_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>, config: &Self::CLRN) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:211:78
[INFO] [stdout]     |
[INFO] [stdout] 211 |     fn sigmoid_pointwise_plain(&self, x: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 211 -     fn sigmoid_pointwise_plain(&self, x: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 211 +     fn sigmoid_pointwise_plain(&self, x: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:573:173
[INFO] [stdout]     |
[INFO] [stdout] 573 | ... &mut SharedTensor<F>, config: &Self::CLRN) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 573 -     fn lrn_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result: &SharedTensor<F>, result_diff: &mut SharedTensor<F>, config: &Self::CLRN) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 573 +     fn lrn_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result: &SharedTensor<F>, result_diff: &mut SharedTensor<F>, config: &Self::CLRN) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:219:107
[INFO] [stdout]     |
[INFO] [stdout] 219 |     fn sigmoid_pointwise_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                                           ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 219 -     fn sigmoid_pointwise_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 219 +     fn sigmoid_pointwise_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:579:106
[INFO] [stdout]     |
[INFO] [stdout] 579 |     fn new_pooling_config(&self, window: &[i32], padding: &[i32], stride: &[i32]) -> Result<Self::CPOOL, ::co::error::Error>;
[INFO] [stdout]     |                                                                                                          ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 579 -     fn new_pooling_config(&self, window: &[i32], padding: &[i32], stride: &[i32]) -> Result<Self::CPOOL, ::co::error::Error>;
[INFO] [stdout] 579 +     fn new_pooling_config(&self, window: &[i32], padding: &[i32], stride: &[i32]) -> Result<Self::CPOOL, co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:228:109
[INFO] [stdout]     |
[INFO] [stdout] 228 |     fn sigmoid_pointwise_grad_plain(&self, x: &SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                                             ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 228 -     fn sigmoid_pointwise_grad_plain(&self, x: &SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 228 +     fn sigmoid_pointwise_grad_plain(&self, x: &SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:587:118
[INFO] [stdout]     |
[INFO] [stdout] 587 |     fn pooling_max(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>, config: &Self::CPOOL) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                                                      ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 587 -     fn pooling_max(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>, config: &Self::CPOOL) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 587 +     fn pooling_max(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>, config: &Self::CPOOL) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:239:89
[INFO] [stdout]     |
[INFO] [stdout] 239 |     fn relu(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                         ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 239 -     fn relu(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 239 +     fn relu(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:596:120
[INFO] [stdout]     |
[INFO] [stdout] 596 |     fn pooling_max_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>, config: &Self::CPOOL) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                                                        ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 596 -     fn pooling_max_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>, config: &Self::CPOOL) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 596 +     fn pooling_max_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>, config: &Self::CPOOL) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:604:188
[INFO] [stdout]     |
[INFO] [stdout] 604 | ...&mut SharedTensor<F>, config: &Self::CPOOL) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 604 -     fn pooling_max_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>, config: &Self::CPOOL) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 604 +     fn pooling_max_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>, config: &Self::CPOOL) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:248:91
[INFO] [stdout]     |
[INFO] [stdout] 248 |     fn relu_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                           ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 248 -     fn relu_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 248 +     fn relu_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:613:182
[INFO] [stdout]     |
[INFO] [stdout] 613 | ...&mut SharedTensor<F>, config: &Self::CPOOL) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 613 -     fn pooling_max_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result: &SharedTensor<F>, result_diff: &mut SharedTensor<F>, config: &Self::CPOOL) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 613 +     fn pooling_max_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result: &SharedTensor<F>, result_diff: &mut SharedTensor<F>, config: &Self::CPOOL) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:256:159
[INFO] [stdout]     |
[INFO] [stdout] 256 | ...nsor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 256 -     fn relu_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 256 +     fn relu_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:265:153
[INFO] [stdout]     |
[INFO] [stdout] 265 | ...nsor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 265 -     fn relu_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result: &SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 265 +     fn relu_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result: &SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:100:58
[INFO] [stdout]     |
[INFO] [stdout] 100 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:28:1
[INFO] [stdout]     |
[INFO] [stdout] 28  | impl_ops_sigmoid_for!(f32, Backend<Native>);
[INFO] [stdout]     | ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: `#[warn(deprecated)]` on by default
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:276:69
[INFO] [stdout]     |
[INFO] [stdout] 276 |     fn relu_pointwise(&self, x: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                     ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 276 -     fn relu_pointwise(&self, x: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 276 +     fn relu_pointwise(&self, x: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:125:58
[INFO] [stdout]     |
[INFO] [stdout] 125 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:28:1
[INFO] [stdout]     |
[INFO] [stdout] 28  | impl_ops_sigmoid_for!(f32, Backend<Native>);
[INFO] [stdout]     | ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:285:75
[INFO] [stdout]     |
[INFO] [stdout] 285 |     fn relu_pointwise_plain(&self, x: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                           ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 285 -     fn relu_pointwise_plain(&self, x: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 285 +     fn relu_pointwise_plain(&self, x: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:126:63
[INFO] [stdout]     |
[INFO] [stdout] 126 |                 match x_diff.add_device(self.device()) { _ => try!(x_diff.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:28:1
[INFO] [stdout]     |
[INFO] [stdout] 28  | impl_ops_sigmoid_for!(f32, Backend<Native>);
[INFO] [stdout]     | ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:293:104
[INFO] [stdout]     |
[INFO] [stdout] 293 |     fn relu_pointwise_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                                        ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 293 -     fn relu_pointwise_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 293 +     fn relu_pointwise_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:127:63
[INFO] [stdout]     |
[INFO] [stdout] 127 |                 match result.add_device(self.device()) { _ => try!(result.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:28:1
[INFO] [stdout]     |
[INFO] [stdout] 28  | impl_ops_sigmoid_for!(f32, Backend<Native>);
[INFO] [stdout]     | ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:302:106
[INFO] [stdout]     |
[INFO] [stdout] 302 |     fn relu_pointwise_grad_plain(&self, x: &SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                                          ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 302 -     fn relu_pointwise_grad_plain(&self, x: &SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 302 +     fn relu_pointwise_grad_plain(&self, x: &SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:163:58
[INFO] [stdout]     |
[INFO] [stdout] 163 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:29:1
[INFO] [stdout]     |
[INFO] [stdout] 29  | impl_ops_relu_for!(f32, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:313:89
[INFO] [stdout]     |
[INFO] [stdout] 313 |     fn tanh(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                         ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 313 -     fn tanh(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 313 +     fn tanh(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:188:58
[INFO] [stdout]     |
[INFO] [stdout] 188 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:29:1
[INFO] [stdout]     |
[INFO] [stdout] 29  | impl_ops_relu_for!(f32, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:189:63
[INFO] [stdout]     |
[INFO] [stdout] 189 |                 match x_diff.add_device(self.device()) { _ => try!(x_diff.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:29:1
[INFO] [stdout]     |
[INFO] [stdout] 29  | impl_ops_relu_for!(f32, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:322:91
[INFO] [stdout]     |
[INFO] [stdout] 322 |     fn tanh_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                           ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 322 -     fn tanh_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 322 +     fn tanh_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:190:63
[INFO] [stdout]     |
[INFO] [stdout] 190 |                 match result.add_device(self.device()) { _ => try!(result.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:29:1
[INFO] [stdout]     |
[INFO] [stdout] 29  | impl_ops_relu_for!(f32, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:330:159
[INFO] [stdout]     |
[INFO] [stdout] 330 | ...nsor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 330 -     fn tanh_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 330 +     fn tanh_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:339:153
[INFO] [stdout]     |
[INFO] [stdout] 339 | ...nsor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 339 -     fn tanh_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result: &SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 339 +     fn tanh_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result: &SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:227:58
[INFO] [stdout]     |
[INFO] [stdout] 227 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:30:1
[INFO] [stdout]     |
[INFO] [stdout] 30  | impl_ops_tanh_for!(f32, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:350:69
[INFO] [stdout]     |
[INFO] [stdout] 350 |     fn tanh_pointwise(&self, x: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                     ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 350 -     fn tanh_pointwise(&self, x: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 350 +     fn tanh_pointwise(&self, x: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:252:58
[INFO] [stdout]     |
[INFO] [stdout] 252 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:30:1
[INFO] [stdout]     |
[INFO] [stdout] 30  | impl_ops_tanh_for!(f32, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:359:75
[INFO] [stdout]     |
[INFO] [stdout] 359 |     fn tanh_pointwise_plain(&self, x: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                           ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 359 -     fn tanh_pointwise_plain(&self, x: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 359 +     fn tanh_pointwise_plain(&self, x: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:253:63
[INFO] [stdout]     |
[INFO] [stdout] 253 |                 match x_diff.add_device(self.device()) { _ => try!(x_diff.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:30:1
[INFO] [stdout]     |
[INFO] [stdout] 30  | impl_ops_tanh_for!(f32, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:367:104
[INFO] [stdout]     |
[INFO] [stdout] 367 |     fn tanh_pointwise_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                                        ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 367 -     fn tanh_pointwise_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 367 +     fn tanh_pointwise_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:254:63
[INFO] [stdout]     |
[INFO] [stdout] 254 |                 match result.add_device(self.device()) { _ => try!(result.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:30:1
[INFO] [stdout]     |
[INFO] [stdout] 30  | impl_ops_tanh_for!(f32, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:376:106
[INFO] [stdout]     |
[INFO] [stdout] 376 |     fn tanh_pointwise_grad_plain(&self, x: &SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                                          ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 376 -     fn tanh_pointwise_grad_plain(&self, x: &SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 376 +     fn tanh_pointwise_grad_plain(&self, x: &SharedTensor<F>, x_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:352:58
[INFO] [stdout]     |
[INFO] [stdout] 352 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:32:2
[INFO] [stdout]     |
[INFO] [stdout] 32  |  impl_ops_softmax_for!(f32, Backend<Native>);
[INFO] [stdout]     |  ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:384:87
[INFO] [stdout]     |
[INFO] [stdout] 384 | ...                   stride: &[i32], zero_padding: &[i32]) -> Result<Self::CC, ::co::error::Error>;
[INFO] [stdout]     |                                                                                 ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 384 -                             stride: &[i32], zero_padding: &[i32]) -> Result<Self::CC, ::co::error::Error>;
[INFO] [stdout] 384 +                             stride: &[i32], zero_padding: &[i32]) -> Result<Self::CC, co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:381:58
[INFO] [stdout]     |
[INFO] [stdout] 381 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:32:2
[INFO] [stdout]     |
[INFO] [stdout] 32  |  impl_ops_softmax_for!(f32, Backend<Native>);
[INFO] [stdout]     |  ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:392:179
[INFO] [stdout]     |
[INFO] [stdout] 392 | ...: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 392 -     fn convolution(&self, filter: &mut SharedTensor<F>, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>, workspace: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 392 +     fn convolution(&self, filter: &mut SharedTensor<F>, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>, workspace: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:382:63
[INFO] [stdout]     |
[INFO] [stdout] 382 |                 match x_diff.add_device(self.device()) { _ => try!(x_diff.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:32:2
[INFO] [stdout]     |
[INFO] [stdout] 32  |  impl_ops_softmax_for!(f32, Backend<Native>);
[INFO] [stdout]     |  ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:401:177
[INFO] [stdout]     |
[INFO] [stdout] 401 | ...: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 401 -     fn convolution_plain(&self, filter: &SharedTensor<F>, x: &SharedTensor<F>, result: &mut SharedTensor<F>, workspace: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 401 +     fn convolution_plain(&self, filter: &SharedTensor<F>, x: &SharedTensor<F>, result: &mut SharedTensor<F>, workspace: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:424:58
[INFO] [stdout]     |
[INFO] [stdout] 424 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:33:2
[INFO] [stdout]     |
[INFO] [stdout] 33  |  impl_ops_log_softmax_for!(f32, Backend<Native>);
[INFO] [stdout]     |  ----------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:409:206
[INFO] [stdout]     |
[INFO] [stdout] 409 | ...: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 409 -     fn convolution_grad_filter(&self, src_data: &mut SharedTensor<F>, dest_diff: &mut SharedTensor<F>, filter_diff: &mut SharedTensor<F>, workspace: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 409 +     fn convolution_grad_filter(&self, src_data: &mut SharedTensor<F>, dest_diff: &mut SharedTensor<F>, filter_diff: &mut SharedTensor<F>, workspace: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:459:58
[INFO] [stdout]     |
[INFO] [stdout] 459 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:33:2
[INFO] [stdout]     |
[INFO] [stdout] 33  |  impl_ops_log_softmax_for!(f32, Backend<Native>);
[INFO] [stdout]     |  ----------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:418:204
[INFO] [stdout]     |
[INFO] [stdout] 418 | ...: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 418 -     fn convolution_grad_filter_plain(&self, src_data: &SharedTensor<F>, dest_diff: &SharedTensor<F>, filter_diff: &mut SharedTensor<F>, workspace: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 418 +     fn convolution_grad_filter_plain(&self, src_data: &SharedTensor<F>, dest_diff: &SharedTensor<F>, filter_diff: &mut SharedTensor<F>, workspace: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:460:63
[INFO] [stdout]     |
[INFO] [stdout] 460 |                 match x_diff.add_device(self.device()) { _ => try!(x_diff.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:33:2
[INFO] [stdout]     |
[INFO] [stdout] 33  |  impl_ops_log_softmax_for!(f32, Backend<Native>);
[INFO] [stdout]     |  ----------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:100:58
[INFO] [stdout]     |
[INFO] [stdout] 100 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:46:1
[INFO] [stdout]     |
[INFO] [stdout] 46  | impl_ops_sigmoid_for!(f64, Backend<Native>);
[INFO] [stdout]     | ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:426:199
[INFO] [stdout]     |
[INFO] [stdout] 426 | ...: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 426 -     fn convolution_grad_data(&self, filter: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>, workspace: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 426 +     fn convolution_grad_data(&self, filter: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>, workspace: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:435:197
[INFO] [stdout]     |
[INFO] [stdout] 435 | ...: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 435 -     fn convolution_grad_data_plain(&self, filter: &SharedTensor<F>, x_diff: &SharedTensor<F>, result_diff: &mut SharedTensor<F>, workspace: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 435 +     fn convolution_grad_data_plain(&self, filter: &SharedTensor<F>, x_diff: &SharedTensor<F>, result_diff: &mut SharedTensor<F>, workspace: &mut SharedTensor<u8>, config: &Self::CC) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:125:58
[INFO] [stdout]     |
[INFO] [stdout] 125 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:46:1
[INFO] [stdout]     |
[INFO] [stdout] 46  | impl_ops_sigmoid_for!(f64, Backend<Native>);
[INFO] [stdout]     | ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:472:92
[INFO] [stdout]     |
[INFO] [stdout] 472 |     fn softmax(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                            ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 472 -     fn softmax(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 472 +     fn softmax(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:126:63
[INFO] [stdout]     |
[INFO] [stdout] 126 |                 match x_diff.add_device(self.device()) { _ => try!(x_diff.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:46:1
[INFO] [stdout]     |
[INFO] [stdout] 46  | impl_ops_sigmoid_for!(f64, Backend<Native>);
[INFO] [stdout]     | ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:481:94
[INFO] [stdout]     |
[INFO] [stdout] 481 |     fn softmax_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 481 -     fn softmax_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 481 +     fn softmax_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:127:63
[INFO] [stdout]     |
[INFO] [stdout] 127 |                 match result.add_device(self.device()) { _ => try!(result.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:46:1
[INFO] [stdout]     |
[INFO] [stdout] 46  | impl_ops_sigmoid_for!(f64, Backend<Native>);
[INFO] [stdout]     | ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:489:132
[INFO] [stdout]     |
[INFO] [stdout] 489 | ...nsor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 489 -     fn softmax_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 489 +     fn softmax_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:163:58
[INFO] [stdout]     |
[INFO] [stdout] 163 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:47:1
[INFO] [stdout]     |
[INFO] [stdout] 47  | impl_ops_relu_for!(f64, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:188:58
[INFO] [stdout]     |
[INFO] [stdout] 188 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:47:1
[INFO] [stdout]     |
[INFO] [stdout] 47  | impl_ops_relu_for!(f64, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:498:130
[INFO] [stdout]     |
[INFO] [stdout] 498 | ...nsor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 498 -     fn softmax_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 498 +     fn softmax_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:189:63
[INFO] [stdout]     |
[INFO] [stdout] 189 |                 match x_diff.add_device(self.device()) { _ => try!(x_diff.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:47:1
[INFO] [stdout]     |
[INFO] [stdout] 47  | impl_ops_relu_for!(f64, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:508:96
[INFO] [stdout]     |
[INFO] [stdout] 508 |     fn log_softmax(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                                ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 508 -     fn log_softmax(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 508 +     fn log_softmax(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:190:63
[INFO] [stdout]     |
[INFO] [stdout] 190 |                 match result.add_device(self.device()) { _ => try!(result.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:47:1
[INFO] [stdout]     |
[INFO] [stdout] 47  | impl_ops_relu_for!(f64, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:517:98
[INFO] [stdout]     |
[INFO] [stdout] 517 |     fn log_softmax_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                                  ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 517 -     fn log_softmax_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 517 +     fn log_softmax_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:227:58
[INFO] [stdout]     |
[INFO] [stdout] 227 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:48:1
[INFO] [stdout]     |
[INFO] [stdout] 48  | impl_ops_tanh_for!(f64, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:524:136
[INFO] [stdout]     |
[INFO] [stdout] 524 | ...nsor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 524 -     fn log_softmax_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 524 +     fn log_softmax_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:252:58
[INFO] [stdout]     |
[INFO] [stdout] 252 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:48:1
[INFO] [stdout]     |
[INFO] [stdout] 48  | impl_ops_tanh_for!(f64, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:533:134
[INFO] [stdout]     |
[INFO] [stdout] 533 | ...nsor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 533 -     fn log_softmax_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 533 +     fn log_softmax_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result_diff: &mut SharedTensor<F>) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:253:63
[INFO] [stdout]     |
[INFO] [stdout] 253 |                 match x_diff.add_device(self.device()) { _ => try!(x_diff.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:48:1
[INFO] [stdout]     |
[INFO] [stdout] 48  | impl_ops_tanh_for!(f64, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:539:91
[INFO] [stdout]     |
[INFO] [stdout] 539 |     fn new_lrn_config(&self, n: u32, alpha: f64, beta: f64, k: f64) -> Result<Self::CLRN, ::co::error::Error>;
[INFO] [stdout]     |                                                                                           ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 539 -     fn new_lrn_config(&self, n: u32, alpha: f64, beta: f64, k: f64) -> Result<Self::CLRN, ::co::error::Error>;
[INFO] [stdout] 539 +     fn new_lrn_config(&self, n: u32, alpha: f64, beta: f64, k: f64) -> Result<Self::CLRN, co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:254:63
[INFO] [stdout]     |
[INFO] [stdout] 254 |                 match result.add_device(self.device()) { _ => try!(result.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:48:1
[INFO] [stdout]     |
[INFO] [stdout] 48  | impl_ops_tanh_for!(f64, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:547:109
[INFO] [stdout]     |
[INFO] [stdout] 547 |     fn lrn(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>, config: &Self::CLRN) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                                             ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 547 -     fn lrn(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>, config: &Self::CLRN) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 547 +     fn lrn(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>, config: &Self::CLRN) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:352:58
[INFO] [stdout]     |
[INFO] [stdout] 352 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:50:2
[INFO] [stdout]     |
[INFO] [stdout] 50  |  impl_ops_softmax_for!(f64, Backend<Native>);
[INFO] [stdout]     |  ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:556:111
[INFO] [stdout]     |
[INFO] [stdout] 556 |     fn lrn_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>, config: &Self::CLRN) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                                               ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 556 -     fn lrn_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>, config: &Self::CLRN) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 556 +     fn lrn_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>, config: &Self::CLRN) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:564:179
[INFO] [stdout]     |
[INFO] [stdout] 564 | ... &mut SharedTensor<F>, config: &Self::CLRN) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 564 -     fn lrn_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>, config: &Self::CLRN) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 564 +     fn lrn_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>, config: &Self::CLRN) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:381:58
[INFO] [stdout]     |
[INFO] [stdout] 381 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:50:2
[INFO] [stdout]     |
[INFO] [stdout] 50  |  impl_ops_softmax_for!(f64, Backend<Native>);
[INFO] [stdout]     |  ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:573:173
[INFO] [stdout]     |
[INFO] [stdout] 573 | ... &mut SharedTensor<F>, config: &Self::CLRN) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 573 -     fn lrn_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result: &SharedTensor<F>, result_diff: &mut SharedTensor<F>, config: &Self::CLRN) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 573 +     fn lrn_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result: &SharedTensor<F>, result_diff: &mut SharedTensor<F>, config: &Self::CLRN) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:579:106
[INFO] [stdout]     |
[INFO] [stdout] 579 |     fn new_pooling_config(&self, window: &[i32], padding: &[i32], stride: &[i32]) -> Result<Self::CPOOL, ::co::error::Error>;
[INFO] [stdout]     |                                                                                                          ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 579 -     fn new_pooling_config(&self, window: &[i32], padding: &[i32], stride: &[i32]) -> Result<Self::CPOOL, ::co::error::Error>;
[INFO] [stdout] 579 +     fn new_pooling_config(&self, window: &[i32], padding: &[i32], stride: &[i32]) -> Result<Self::CPOOL, co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:382:63
[INFO] [stdout]     |
[INFO] [stdout] 382 |                 match x_diff.add_device(self.device()) { _ => try!(x_diff.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:50:2
[INFO] [stdout]     |
[INFO] [stdout] 50  |  impl_ops_softmax_for!(f64, Backend<Native>);
[INFO] [stdout]     |  ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:587:118
[INFO] [stdout]     |
[INFO] [stdout] 587 |     fn pooling_max(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>, config: &Self::CPOOL) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                                                      ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 587 -     fn pooling_max(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>, config: &Self::CPOOL) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 587 +     fn pooling_max(&self, x: &mut SharedTensor<F>, result: &mut SharedTensor<F>, config: &Self::CPOOL) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:424:58
[INFO] [stdout]     |
[INFO] [stdout] 424 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:51:2
[INFO] [stdout]     |
[INFO] [stdout] 51  |  impl_ops_log_softmax_for!(f64, Backend<Native>);
[INFO] [stdout]     |  ----------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:459:58
[INFO] [stdout]     |
[INFO] [stdout] 459 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:51:2
[INFO] [stdout]     |
[INFO] [stdout] 51  |  impl_ops_log_softmax_for!(f64, Backend<Native>);
[INFO] [stdout]     |  ----------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:596:120
[INFO] [stdout]     |
[INFO] [stdout] 596 |     fn pooling_max_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>, config: &Self::CPOOL) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                                                                                        ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 596 -     fn pooling_max_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>, config: &Self::CPOOL) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 596 +     fn pooling_max_plain(&self, x: &SharedTensor<F>, result: &mut SharedTensor<F>, config: &Self::CPOOL) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:460:63
[INFO] [stdout]     |
[INFO] [stdout] 460 |                 match x_diff.add_device(self.device()) { _ => try!(x_diff.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:51:2
[INFO] [stdout]     |
[INFO] [stdout] 51  |  impl_ops_log_softmax_for!(f64, Backend<Native>);
[INFO] [stdout]     |  ----------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:604:188
[INFO] [stdout]     |
[INFO] [stdout] 604 | ...&mut SharedTensor<F>, config: &Self::CPOOL) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 604 -     fn pooling_max_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>, config: &Self::CPOOL) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 604 +     fn pooling_max_grad(&self, x: &mut SharedTensor<F>, x_diff: &mut SharedTensor<F>, result: &mut SharedTensor<F>, result_diff: &mut SharedTensor<F>, config: &Self::CPOOL) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]  --> src/frameworks/cuda/helper.rs:4:41
[INFO] [stdout]   |
[INFO] [stdout] 4 | pub unsafe fn receive_memory_ptr<T>(x: &::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*const ::libc::c_void...
[INFO] [stdout]   |                                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]   |
[INFO] [stdout] 4 - pub unsafe fn receive_memory_ptr<T>(x: &::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*const ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout] 4 + pub unsafe fn receive_memory_ptr<T>(x: &co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*const ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout]   |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/plugin.rs:613:182
[INFO] [stdout]     |
[INFO] [stdout] 613 | ...&mut SharedTensor<F>, config: &Self::CPOOL) -> Result<(), ::co::error::Error>;
[INFO] [stdout]     |                                                              ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 613 -     fn pooling_max_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result: &SharedTensor<F>, result_diff: &mut SharedTensor<F>, config: &Self::CPOOL) -> Result<(), ::co::error::Error>;
[INFO] [stdout] 613 +     fn pooling_max_grad_plain(&self, x: &SharedTensor<F>, x_diff: &SharedTensor<F>, result: &SharedTensor<F>, result_diff: &mut SharedTensor<F>, config: &Self::CPOOL) -> Result<(), co::error::Error>;
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]  --> src/frameworks/cuda/helper.rs:4:81
[INFO] [stdout]   |
[INFO] [stdout] 4 | pub unsafe fn receive_memory_ptr<T>(x: &::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*const ::libc::c_void...
[INFO] [stdout]   |                                                                                 ^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]   |
[INFO] [stdout] 4 - pub unsafe fn receive_memory_ptr<T>(x: &::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*const ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout] 4 + pub unsafe fn receive_memory_ptr<T>(x: &::co::tensor::SharedTensor<T>, device: &co::device::DeviceType) -> Result<*const ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout]   |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:100:58
[INFO] [stdout]     |
[INFO] [stdout] 100 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:28:1
[INFO] [stdout]     |
[INFO] [stdout] 28  | impl_ops_sigmoid_for!(f32, Backend<Native>);
[INFO] [stdout]     | ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: `#[warn(deprecated)]` on by default
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]  --> src/frameworks/cuda/helper.rs:4:124
[INFO] [stdout]   |
[INFO] [stdout] 4 | ...sor<T>, device: &::co::device::DeviceType) -> Result<*const ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout]   |                                                                ^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]   |
[INFO] [stdout] 4 - pub unsafe fn receive_memory_ptr<T>(x: &::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*const ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout] 4 + pub unsafe fn receive_memory_ptr<T>(x: &::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*const libc::c_void, ::co::plugin::Error> {
[INFO] [stdout]   |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:125:58
[INFO] [stdout]     |
[INFO] [stdout] 125 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:28:1
[INFO] [stdout]     |
[INFO] [stdout] 28  | impl_ops_sigmoid_for!(f32, Backend<Native>);
[INFO] [stdout]     | ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]  --> src/frameworks/cuda/helper.rs:4:140
[INFO] [stdout]   |
[INFO] [stdout] 4 | ...co::device::DeviceType) -> Result<*const ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout]   |                                                             ^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]   |
[INFO] [stdout] 4 - pub unsafe fn receive_memory_ptr<T>(x: &::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*const ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout] 4 + pub unsafe fn receive_memory_ptr<T>(x: &::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*const ::libc::c_void, co::plugin::Error> {
[INFO] [stdout]   |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:126:63
[INFO] [stdout]     |
[INFO] [stdout] 126 |                 match x_diff.add_device(self.device()) { _ => try!(x_diff.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:28:1
[INFO] [stdout]     |
[INFO] [stdout] 28  | impl_ops_sigmoid_for!(f32, Backend<Native>);
[INFO] [stdout]     | ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]  --> src/frameworks/cuda/helper.rs:5:8
[INFO] [stdout]   |
[INFO] [stdout] 5 |     Ok(::std::mem::transmute::<u64, *const ::libc::c_void>(
[INFO] [stdout]   |        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]   |
[INFO] [stdout] 5 -     Ok(::std::mem::transmute::<u64, *const ::libc::c_void>(
[INFO] [stdout] 5 +     Ok(std::mem::transmute::<u64, *const ::libc::c_void>(
[INFO] [stdout]   |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:127:63
[INFO] [stdout]     |
[INFO] [stdout] 127 |                 match result.add_device(self.device()) { _ => try!(result.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:28:1
[INFO] [stdout]     |
[INFO] [stdout] 28  | impl_ops_sigmoid_for!(f32, Backend<Native>);
[INFO] [stdout]     | ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]  --> src/frameworks/cuda/helper.rs:5:44
[INFO] [stdout]   |
[INFO] [stdout] 5 |     Ok(::std::mem::transmute::<u64, *const ::libc::c_void>(
[INFO] [stdout]   |                                            ^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]   |
[INFO] [stdout] 5 -     Ok(::std::mem::transmute::<u64, *const ::libc::c_void>(
[INFO] [stdout] 5 +     Ok(::std::mem::transmute::<u64, *const libc::c_void>(
[INFO] [stdout]   |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:163:58
[INFO] [stdout]     |
[INFO] [stdout] 163 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:29:1
[INFO] [stdout]     |
[INFO] [stdout] 29  | impl_ops_relu_for!(f32, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]  --> src/frameworks/cuda/helper.rs:6:10
[INFO] [stdout]   |
[INFO] [stdout] 6 |         *try!(
[INFO] [stdout]   |          ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]  --> src/frameworks/cuda/helper.rs:7:13
[INFO] [stdout]   |
[INFO] [stdout] 7 |             try!(
[INFO] [stdout]   |             ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:188:58
[INFO] [stdout]     |
[INFO] [stdout] 188 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:29:1
[INFO] [stdout]     |
[INFO] [stdout] 29  | impl_ops_relu_for!(f32, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]  --> src/frameworks/cuda/helper.rs:8:37
[INFO] [stdout]   |
[INFO] [stdout] 8 |                 x.get(device).ok_or(::co::plugin::Error::MissingMemoryForDevice("Unable to resolve memory."))
[INFO] [stdout]   |                                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]   |
[INFO] [stdout] 8 -                 x.get(device).ok_or(::co::plugin::Error::MissingMemoryForDevice("Unable to resolve memory."))
[INFO] [stdout] 8 +                 x.get(device).ok_or(co::plugin::Error::MissingMemoryForDevice("Unable to resolve memory."))
[INFO] [stdout]   |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:189:63
[INFO] [stdout]     |
[INFO] [stdout] 189 |                 match x_diff.add_device(self.device()) { _ => try!(x_diff.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:29:1
[INFO] [stdout]     |
[INFO] [stdout] 29  | impl_ops_relu_for!(f32, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]  --> src/frameworks/cuda/helper.rs:9:31
[INFO] [stdout]   |
[INFO] [stdout] 9 |             ).as_cuda().ok_or(::co::plugin::Error::MissingMemoryForDevice("Unable to receive CUDA memory."))
[INFO] [stdout]   |                               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]   |
[INFO] [stdout] 9 -             ).as_cuda().ok_or(::co::plugin::Error::MissingMemoryForDevice("Unable to receive CUDA memory."))
[INFO] [stdout] 9 +             ).as_cuda().ok_or(co::plugin::Error::MissingMemoryForDevice("Unable to receive CUDA memory."))
[INFO] [stdout]   |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:190:63
[INFO] [stdout]     |
[INFO] [stdout] 190 |                 match result.add_device(self.device()) { _ => try!(result.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:29:1
[INFO] [stdout]     |
[INFO] [stdout] 29  | impl_ops_relu_for!(f32, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]   --> src/frameworks/cuda/helper.rs:15:49
[INFO] [stdout]    |
[INFO] [stdout] 15 | pub unsafe fn receive_memory_ptr_mut<T>(x: &mut ::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*mut ::libc::...
[INFO] [stdout]    |                                                 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]    |
[INFO] [stdout] 15 - pub unsafe fn receive_memory_ptr_mut<T>(x: &mut ::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*mut ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout] 15 + pub unsafe fn receive_memory_ptr_mut<T>(x: &mut co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*mut ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout]    |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:227:58
[INFO] [stdout]     |
[INFO] [stdout] 227 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:30:1
[INFO] [stdout]     |
[INFO] [stdout] 30  | impl_ops_tanh_for!(f32, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:252:58
[INFO] [stdout]     |
[INFO] [stdout] 252 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:30:1
[INFO] [stdout]     |
[INFO] [stdout] 30  | impl_ops_tanh_for!(f32, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]   --> src/frameworks/cuda/helper.rs:15:89
[INFO] [stdout]    |
[INFO] [stdout] 15 | pub unsafe fn receive_memory_ptr_mut<T>(x: &mut ::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*mut ::libc::...
[INFO] [stdout]    |                                                                                         ^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]    |
[INFO] [stdout] 15 - pub unsafe fn receive_memory_ptr_mut<T>(x: &mut ::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*mut ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout] 15 + pub unsafe fn receive_memory_ptr_mut<T>(x: &mut ::co::tensor::SharedTensor<T>, device: &co::device::DeviceType) -> Result<*mut ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout]    |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]   --> src/frameworks/cuda/helper.rs:15:130
[INFO] [stdout]    |
[INFO] [stdout] 15 | ...ensor<T>, device: &::co::device::DeviceType) -> Result<*mut ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout]    |                                                                ^^^^^^^^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]    |
[INFO] [stdout] 15 - pub unsafe fn receive_memory_ptr_mut<T>(x: &mut ::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*mut ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout] 15 + pub unsafe fn receive_memory_ptr_mut<T>(x: &mut ::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*mut libc::c_void, ::co::plugin::Error> {
[INFO] [stdout]    |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:253:63
[INFO] [stdout]     |
[INFO] [stdout] 253 |                 match x_diff.add_device(self.device()) { _ => try!(x_diff.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:30:1
[INFO] [stdout]     |
[INFO] [stdout] 30  | impl_ops_tanh_for!(f32, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]   --> src/frameworks/cuda/helper.rs:15:146
[INFO] [stdout]    |
[INFO] [stdout] 15 | ...::co::device::DeviceType) -> Result<*mut ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout]    |                                                             ^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]    |
[INFO] [stdout] 15 - pub unsafe fn receive_memory_ptr_mut<T>(x: &mut ::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*mut ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout] 15 + pub unsafe fn receive_memory_ptr_mut<T>(x: &mut ::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*mut ::libc::c_void, co::plugin::Error> {
[INFO] [stdout]    |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:254:63
[INFO] [stdout]     |
[INFO] [stdout] 254 |                 match result.add_device(self.device()) { _ => try!(result.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:30:1
[INFO] [stdout]     |
[INFO] [stdout] 30  | impl_ops_tanh_for!(f32, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]   --> src/frameworks/cuda/helper.rs:16:8
[INFO] [stdout]    |
[INFO] [stdout] 16 |     Ok(::std::mem::transmute::<u64, *mut ::libc::c_void>(
[INFO] [stdout]    |        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]    |
[INFO] [stdout] 16 -     Ok(::std::mem::transmute::<u64, *mut ::libc::c_void>(
[INFO] [stdout] 16 +     Ok(std::mem::transmute::<u64, *mut ::libc::c_void>(
[INFO] [stdout]    |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:352:58
[INFO] [stdout]     |
[INFO] [stdout] 352 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:32:2
[INFO] [stdout]     |
[INFO] [stdout] 32  |  impl_ops_softmax_for!(f32, Backend<Native>);
[INFO] [stdout]     |  ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]   --> src/frameworks/cuda/helper.rs:16:42
[INFO] [stdout]    |
[INFO] [stdout] 16 |     Ok(::std::mem::transmute::<u64, *mut ::libc::c_void>(
[INFO] [stdout]    |                                          ^^^^^^^^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]    |
[INFO] [stdout] 16 -     Ok(::std::mem::transmute::<u64, *mut ::libc::c_void>(
[INFO] [stdout] 16 +     Ok(::std::mem::transmute::<u64, *mut libc::c_void>(
[INFO] [stdout]    |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]   --> src/frameworks/cuda/helper.rs:17:10
[INFO] [stdout]    |
[INFO] [stdout] 17 |         *try!(
[INFO] [stdout]    |          ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:381:58
[INFO] [stdout]     |
[INFO] [stdout] 381 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:32:2
[INFO] [stdout]     |
[INFO] [stdout] 32  |  impl_ops_softmax_for!(f32, Backend<Native>);
[INFO] [stdout]     |  ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]   --> src/frameworks/cuda/helper.rs:18:13
[INFO] [stdout]    |
[INFO] [stdout] 18 |             try!(
[INFO] [stdout]    |             ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:382:63
[INFO] [stdout]     |
[INFO] [stdout] 382 |                 match x_diff.add_device(self.device()) { _ => try!(x_diff.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:32:2
[INFO] [stdout]     |
[INFO] [stdout] 32  |  impl_ops_softmax_for!(f32, Backend<Native>);
[INFO] [stdout]     |  ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]   --> src/frameworks/cuda/helper.rs:19:41
[INFO] [stdout]    |
[INFO] [stdout] 19 |                 x.get_mut(device).ok_or(::co::plugin::Error::MissingMemoryForDevice("Unable to resolve memory."))
[INFO] [stdout]    |                                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]    |
[INFO] [stdout] 19 -                 x.get_mut(device).ok_or(::co::plugin::Error::MissingMemoryForDevice("Unable to resolve memory."))
[INFO] [stdout] 19 +                 x.get_mut(device).ok_or(co::plugin::Error::MissingMemoryForDevice("Unable to resolve memory."))
[INFO] [stdout]    |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:424:58
[INFO] [stdout]     |
[INFO] [stdout] 424 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:33:2
[INFO] [stdout]     |
[INFO] [stdout] 33  |  impl_ops_log_softmax_for!(f32, Backend<Native>);
[INFO] [stdout]     |  ----------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]   --> src/frameworks/cuda/helper.rs:20:35
[INFO] [stdout]    |
[INFO] [stdout] 20 |             ).as_mut_cuda().ok_or(::co::plugin::Error::MissingMemoryForDevice("Unable to receive CUDA memory."))
[INFO] [stdout]    |                                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]    |
[INFO] [stdout] 20 -             ).as_mut_cuda().ok_or(::co::plugin::Error::MissingMemoryForDevice("Unable to receive CUDA memory."))
[INFO] [stdout] 20 +             ).as_mut_cuda().ok_or(co::plugin::Error::MissingMemoryForDevice("Unable to receive CUDA memory."))
[INFO] [stdout]    |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:106:40
[INFO] [stdout]     |
[INFO] [stdout] 106 | impl_icudnndesc_for_sharedtensor!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     |                                        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 106 - impl_icudnndesc_for_sharedtensor!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout] 106 + impl_icudnndesc_for_sharedtensor!(f32, utils::DataType::Float);
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:459:58
[INFO] [stdout]     |
[INFO] [stdout] 459 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:33:2
[INFO] [stdout]     |
[INFO] [stdout] 33  |  impl_ops_log_softmax_for!(f32, Backend<Native>);
[INFO] [stdout]     |  ----------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:107:40
[INFO] [stdout]     |
[INFO] [stdout] 107 | impl_icudnndesc_for_sharedtensor!(f64, ::cudnn::utils::DataType::Double);
[INFO] [stdout]     |                                        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 107 - impl_icudnndesc_for_sharedtensor!(f64, ::cudnn::utils::DataType::Double);
[INFO] [stdout] 107 + impl_icudnndesc_for_sharedtensor!(f64, utils::DataType::Double);
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:460:63
[INFO] [stdout]     |
[INFO] [stdout] 460 |                 match x_diff.add_device(self.device()) { _ => try!(x_diff.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:33:2
[INFO] [stdout]     |
[INFO] [stdout] 33  |  impl_ops_log_softmax_for!(f32, Backend<Native>);
[INFO] [stdout]     |  ----------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:100:58
[INFO] [stdout]     |
[INFO] [stdout] 100 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:46:1
[INFO] [stdout]     |
[INFO] [stdout] 46  | impl_ops_sigmoid_for!(f64, Backend<Native>);
[INFO] [stdout]     | ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:115:61
[INFO] [stdout]     |
[INFO] [stdout] 115 |     fn as_cudnn(&self) -> Result<cudnnConvolutionFwdAlgo_t, ::co::error::Error> {
[INFO] [stdout]     |                                                             ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 115 -     fn as_cudnn(&self) -> Result<cudnnConvolutionFwdAlgo_t, ::co::error::Error> {
[INFO] [stdout] 115 +     fn as_cudnn(&self) -> Result<cudnnConvolutionFwdAlgo_t, co::error::Error> {
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:125:58
[INFO] [stdout]     |
[INFO] [stdout] 125 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:46:1
[INFO] [stdout]     |
[INFO] [stdout] 46  | impl_ops_sigmoid_for!(f64, Backend<Native>);
[INFO] [stdout]     | ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:117:49
[INFO] [stdout]     |
[INFO] [stdout] 117 | ...   ConvForwardAlgo::Auto => return Err(::co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution forward ...
[INFO] [stdout]     |                                           ^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 117 -             ConvForwardAlgo::Auto => return Err(::co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution forward algorithm from ConvForwardAlgo::Auto. Use `find_cudnn_algo` to find an algorithm."))),
[INFO] [stdout] 117 +             ConvForwardAlgo::Auto => return Err(co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution forward algorithm from ConvForwardAlgo::Auto. Use `find_cudnn_algo` to find an algorithm."))),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:126:63
[INFO] [stdout]     |
[INFO] [stdout] 126 |                 match x_diff.add_device(self.device()) { _ => try!(x_diff.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:46:1
[INFO] [stdout]     |
[INFO] [stdout] 46  | impl_ops_sigmoid_for!(f64, Backend<Native>);
[INFO] [stdout]     | ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:117:76
[INFO] [stdout]     |
[INFO] [stdout] 117 | ...   ConvForwardAlgo::Auto => return Err(::co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution forward ...
[INFO] [stdout]     |                                                                      ^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 117 -             ConvForwardAlgo::Auto => return Err(::co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution forward algorithm from ConvForwardAlgo::Auto. Use `find_cudnn_algo` to find an algorithm."))),
[INFO] [stdout] 117 +             ConvForwardAlgo::Auto => return Err(::co::error::Error::Plugin(co::plugin::Error::Plugin("Can't create cuDNN convolution forward algorithm from ConvForwardAlgo::Auto. Use `find_cudnn_algo` to find an algorithm."))),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:127:63
[INFO] [stdout]     |
[INFO] [stdout] 127 |                 match result.add_device(self.device()) { _ => try!(result.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:46:1
[INFO] [stdout]     |
[INFO] [stdout] 46  | impl_ops_sigmoid_for!(f64, Backend<Native>);
[INFO] [stdout]     | ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:118:38
[INFO] [stdout]     |
[INFO] [stdout] 118 |             ConvForwardAlgo::GEMM => ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_GEMM,
[INFO] [stdout]     |                                      ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 118 -             ConvForwardAlgo::GEMM => ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_GEMM,
[INFO] [stdout] 118 +             ConvForwardAlgo::GEMM => cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_GEMM,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:163:58
[INFO] [stdout]     |
[INFO] [stdout] 163 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:47:1
[INFO] [stdout]     |
[INFO] [stdout] 47  | impl_ops_relu_for!(f64, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:119:46
[INFO] [stdout]     |
[INFO] [stdout] 119 |             ConvForwardAlgo::ImplicitGEMM => ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM,
[INFO] [stdout]     |                                              ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 119 -             ConvForwardAlgo::ImplicitGEMM => ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM,
[INFO] [stdout] 119 +             ConvForwardAlgo::ImplicitGEMM => cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:188:58
[INFO] [stdout]     |
[INFO] [stdout] 188 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:47:1
[INFO] [stdout]     |
[INFO] [stdout] 47  | impl_ops_relu_for!(f64, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:120:57
[INFO] [stdout]     |
[INFO] [stdout] 120 | ...   ConvForwardAlgo::ImplicitPrecompiledGEMM => ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_PRECOMP_GEMM,
[INFO] [stdout]     |                                                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 120 -             ConvForwardAlgo::ImplicitPrecompiledGEMM => ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_PRECOMP_GEMM,
[INFO] [stdout] 120 +             ConvForwardAlgo::ImplicitPrecompiledGEMM => cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_PRECOMP_GEMM,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:189:63
[INFO] [stdout]     |
[INFO] [stdout] 189 |                 match x_diff.add_device(self.device()) { _ => try!(x_diff.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:47:1
[INFO] [stdout]     |
[INFO] [stdout] 47  | impl_ops_relu_for!(f64, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:121:37
[INFO] [stdout]     |
[INFO] [stdout] 121 |             ConvForwardAlgo::FFT => ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_FFT,
[INFO] [stdout]     |                                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 121 -             ConvForwardAlgo::FFT => ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_FFT,
[INFO] [stdout] 121 +             ConvForwardAlgo::FFT => cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_FFT,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:190:63
[INFO] [stdout]     |
[INFO] [stdout] 190 |                 match result.add_device(self.device()) { _ => try!(result.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:47:1
[INFO] [stdout]     |
[INFO] [stdout] 47  | impl_ops_relu_for!(f64, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:122:43
[INFO] [stdout]     |
[INFO] [stdout] 122 |             ConvForwardAlgo::FFTTiling => ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_FFT_TILING,
[INFO] [stdout]     |                                           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 122 -             ConvForwardAlgo::FFTTiling => ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_FFT_TILING,
[INFO] [stdout] 122 +             ConvForwardAlgo::FFTTiling => cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_FFT_TILING,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:227:58
[INFO] [stdout]     |
[INFO] [stdout] 227 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:48:1
[INFO] [stdout]     |
[INFO] [stdout] 48  | impl_ops_tanh_for!(f64, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:123:40
[INFO] [stdout]     |
[INFO] [stdout] 123 |             ConvForwardAlgo::Direct => ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_DIRECT,
[INFO] [stdout]     |                                        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 123 -             ConvForwardAlgo::Direct => ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_DIRECT,
[INFO] [stdout] 123 +             ConvForwardAlgo::Direct => cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_DIRECT,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:252:58
[INFO] [stdout]     |
[INFO] [stdout] 252 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:48:1
[INFO] [stdout]     |
[INFO] [stdout] 48  | impl_ops_tanh_for!(f64, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:130:13
[INFO] [stdout]     |
[INFO] [stdout] 130 |             ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_GEMM => ConvForwardAlgo::GEMM,
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 130 -             ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_GEMM => ConvForwardAlgo::GEMM,
[INFO] [stdout] 130 +             cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_GEMM => ConvForwardAlgo::GEMM,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:253:63
[INFO] [stdout]     |
[INFO] [stdout] 253 |                 match x_diff.add_device(self.device()) { _ => try!(x_diff.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:48:1
[INFO] [stdout]     |
[INFO] [stdout] 48  | impl_ops_tanh_for!(f64, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:131:13
[INFO] [stdout]     |
[INFO] [stdout] 131 |             ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM => ConvForwardAlgo::ImplicitGEMM,
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 131 -             ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM => ConvForwardAlgo::ImplicitGEMM,
[INFO] [stdout] 131 +             cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM => ConvForwardAlgo::ImplicitGEMM,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:254:63
[INFO] [stdout]     |
[INFO] [stdout] 254 |                 match result.add_device(self.device()) { _ => try!(result.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:48:1
[INFO] [stdout]     |
[INFO] [stdout] 48  | impl_ops_tanh_for!(f64, Backend<Native>);
[INFO] [stdout]     | ---------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:132:13
[INFO] [stdout]     |
[INFO] [stdout] 132 | ...   ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_PRECOMP_GEMM => ConvForwardAlgo::ImplicitPrecompiledGEMM,
[INFO] [stdout]     |       ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 132 -             ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_PRECOMP_GEMM => ConvForwardAlgo::ImplicitPrecompiledGEMM,
[INFO] [stdout] 132 +             cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_PRECOMP_GEMM => ConvForwardAlgo::ImplicitPrecompiledGEMM,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:133:13
[INFO] [stdout]     |
[INFO] [stdout] 133 |             ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_FFT => ConvForwardAlgo::FFT,
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 133 -             ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_FFT => ConvForwardAlgo::FFT,
[INFO] [stdout] 133 +             cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_FFT => ConvForwardAlgo::FFT,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:352:58
[INFO] [stdout]     |
[INFO] [stdout] 352 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:50:2
[INFO] [stdout]     |
[INFO] [stdout] 50  |  impl_ops_softmax_for!(f64, Backend<Native>);
[INFO] [stdout]     |  ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:134:13
[INFO] [stdout]     |
[INFO] [stdout] 134 |             ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_FFT_TILING => ConvForwardAlgo::FFTTiling,
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 134 -             ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_FFT_TILING => ConvForwardAlgo::FFTTiling,
[INFO] [stdout] 134 +             cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_FFT_TILING => ConvForwardAlgo::FFTTiling,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:381:58
[INFO] [stdout]     |
[INFO] [stdout] 381 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:50:2
[INFO] [stdout]     |
[INFO] [stdout] 50  |  impl_ops_softmax_for!(f64, Backend<Native>);
[INFO] [stdout]     |  ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:135:13
[INFO] [stdout]     |
[INFO] [stdout] 135 |             ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_DIRECT => ConvForwardAlgo::Direct,
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 135 -             ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_DIRECT => ConvForwardAlgo::Direct,
[INFO] [stdout] 135 +             cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_DIRECT => ConvForwardAlgo::Direct,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:382:63
[INFO] [stdout]     |
[INFO] [stdout] 382 |                 match x_diff.add_device(self.device()) { _ => try!(x_diff.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:50:2
[INFO] [stdout]     |
[INFO] [stdout] 50  |  impl_ops_softmax_for!(f64, Backend<Native>);
[INFO] [stdout]     |  ------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:146:34
[INFO] [stdout]     |
[INFO] [stdout] 146 |     ) -> Result<ConvForwardAlgo, ::co::error::Error> {
[INFO] [stdout]     |                                  ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 146 -     ) -> Result<ConvForwardAlgo, ::co::error::Error> {
[INFO] [stdout] 146 +     ) -> Result<ConvForwardAlgo, co::error::Error> {
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:152:29
[INFO] [stdout]     |
[INFO] [stdout] 152 | ...   0 => return Err(::co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution forward algorithm....
[INFO] [stdout]     |                       ^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 152 -             0 => return Err(::co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution forward algorithm."))),
[INFO] [stdout] 152 +             0 => return Err(co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution forward algorithm."))),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:424:58
[INFO] [stdout]     |
[INFO] [stdout] 424 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:51:2
[INFO] [stdout]     |
[INFO] [stdout] 51  |  impl_ops_log_softmax_for!(f64, Backend<Native>);
[INFO] [stdout]     |  ----------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:459:58
[INFO] [stdout]     |
[INFO] [stdout] 459 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:51:2
[INFO] [stdout]     |
[INFO] [stdout] 51  |  impl_ops_log_softmax_for!(f64, Backend<Native>);
[INFO] [stdout]     |  ----------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:152:56
[INFO] [stdout]     |
[INFO] [stdout] 152 | ...   0 => return Err(::co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution forward algorithm....
[INFO] [stdout]     |                                                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 152 -             0 => return Err(::co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution forward algorithm."))),
[INFO] [stdout] 152 +             0 => return Err(::co::error::Error::Plugin(co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution forward algorithm."))),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/native/helper.rs:460:63
[INFO] [stdout]     |
[INFO] [stdout] 460 |                 match x_diff.add_device(self.device()) { _ => try!(x_diff.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/native/mod.rs:51:2
[INFO] [stdout]     |
[INFO] [stdout] 51  |  impl_ops_log_softmax_for!(f64, Backend<Native>);
[INFO] [stdout]     |  ----------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:161:67
[INFO] [stdout]     |
[INFO] [stdout] 161 |     fn as_cudnn(&self) -> Result<cudnnConvolutionBwdFilterAlgo_t, ::co::error::Error> {
[INFO] [stdout]     |                                                                   ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 161 -     fn as_cudnn(&self) -> Result<cudnnConvolutionBwdFilterAlgo_t, ::co::error::Error> {
[INFO] [stdout] 161 +     fn as_cudnn(&self) -> Result<cudnnConvolutionBwdFilterAlgo_t, co::error::Error> {
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]  --> src/frameworks/cuda/helper.rs:4:41
[INFO] [stdout]   |
[INFO] [stdout] 4 | pub unsafe fn receive_memory_ptr<T>(x: &::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*const ::libc::c_void...
[INFO] [stdout]   |                                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]   |
[INFO] [stdout] 4 - pub unsafe fn receive_memory_ptr<T>(x: &::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*const ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout] 4 + pub unsafe fn receive_memory_ptr<T>(x: &co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*const ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout]   |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]  --> src/frameworks/cuda/helper.rs:4:81
[INFO] [stdout]   |
[INFO] [stdout] 4 | pub unsafe fn receive_memory_ptr<T>(x: &::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*const ::libc::c_void...
[INFO] [stdout]   |                                                                                 ^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]   |
[INFO] [stdout] 4 - pub unsafe fn receive_memory_ptr<T>(x: &::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*const ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout] 4 + pub unsafe fn receive_memory_ptr<T>(x: &::co::tensor::SharedTensor<T>, device: &co::device::DeviceType) -> Result<*const ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout]   |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:163:56
[INFO] [stdout]     |
[INFO] [stdout] 163 | ...   ConvBackwardFilterAlgo::Auto => return Err(::co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution b...
[INFO] [stdout]     |                                                  ^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 163 -             ConvBackwardFilterAlgo::Auto => return Err(::co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution backward filter algorithm from ConvBackwardFilterAlgo::Auto. Use `find_cudnn_algo` to find an algorithm."))),
[INFO] [stdout] 163 +             ConvBackwardFilterAlgo::Auto => return Err(co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution backward filter algorithm from ConvBackwardFilterAlgo::Auto. Use `find_cudnn_algo` to find an algorithm."))),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]  --> src/frameworks/cuda/helper.rs:4:124
[INFO] [stdout]   |
[INFO] [stdout] 4 | ...sor<T>, device: &::co::device::DeviceType) -> Result<*const ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout]   |                                                                ^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]   |
[INFO] [stdout] 4 - pub unsafe fn receive_memory_ptr<T>(x: &::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*const ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout] 4 + pub unsafe fn receive_memory_ptr<T>(x: &::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*const libc::c_void, ::co::plugin::Error> {
[INFO] [stdout]   |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:163:83
[INFO] [stdout]     |
[INFO] [stdout] 163 | ...   ConvBackwardFilterAlgo::Auto => return Err(::co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution b...
[INFO] [stdout]     |                                                                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 163 -             ConvBackwardFilterAlgo::Auto => return Err(::co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution backward filter algorithm from ConvBackwardFilterAlgo::Auto. Use `find_cudnn_algo` to find an algorithm."))),
[INFO] [stdout] 163 +             ConvBackwardFilterAlgo::Auto => return Err(::co::error::Error::Plugin(co::plugin::Error::Plugin("Can't create cuDNN convolution backward filter algorithm from ConvBackwardFilterAlgo::Auto. Use `find_cudnn_algo` to find an algorithm."))),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]  --> src/frameworks/cuda/helper.rs:4:140
[INFO] [stdout]   |
[INFO] [stdout] 4 | ...co::device::DeviceType) -> Result<*const ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout]   |                                                             ^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]   |
[INFO] [stdout] 4 - pub unsafe fn receive_memory_ptr<T>(x: &::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*const ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout] 4 + pub unsafe fn receive_memory_ptr<T>(x: &::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*const ::libc::c_void, co::plugin::Error> {
[INFO] [stdout]   |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:164:53
[INFO] [stdout]     |
[INFO] [stdout] 164 |             ConvBackwardFilterAlgo::ImplicitGEMM => ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_1,
[INFO] [stdout]     |                                                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 164 -             ConvBackwardFilterAlgo::ImplicitGEMM => ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_1,
[INFO] [stdout] 164 +             ConvBackwardFilterAlgo::ImplicitGEMM => cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_1,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]  --> src/frameworks/cuda/helper.rs:5:8
[INFO] [stdout]   |
[INFO] [stdout] 5 |     Ok(::std::mem::transmute::<u64, *const ::libc::c_void>(
[INFO] [stdout]   |        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]   |
[INFO] [stdout] 5 -     Ok(::std::mem::transmute::<u64, *const ::libc::c_void>(
[INFO] [stdout] 5 +     Ok(std::mem::transmute::<u64, *const ::libc::c_void>(
[INFO] [stdout]   |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:165:56
[INFO] [stdout]     |
[INFO] [stdout] 165 |             ConvBackwardFilterAlgo::ImplicitGEMMSum => ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_0,
[INFO] [stdout]     |                                                        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 165 -             ConvBackwardFilterAlgo::ImplicitGEMMSum => ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_0,
[INFO] [stdout] 165 +             ConvBackwardFilterAlgo::ImplicitGEMMSum => cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_0,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]  --> src/frameworks/cuda/helper.rs:5:44
[INFO] [stdout]   |
[INFO] [stdout] 5 |     Ok(::std::mem::transmute::<u64, *const ::libc::c_void>(
[INFO] [stdout]   |                                            ^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]   |
[INFO] [stdout] 5 -     Ok(::std::mem::transmute::<u64, *const ::libc::c_void>(
[INFO] [stdout] 5 +     Ok(::std::mem::transmute::<u64, *const libc::c_void>(
[INFO] [stdout]   |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:166:67
[INFO] [stdout]     |
[INFO] [stdout] 166 | ...plicitPrecompiledGEMMSum => ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_3,
[INFO] [stdout]     |                                ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 166 -             ConvBackwardFilterAlgo::ImplicitPrecompiledGEMMSum => ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_3,
[INFO] [stdout] 166 +             ConvBackwardFilterAlgo::ImplicitPrecompiledGEMMSum => cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_3,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]  --> src/frameworks/cuda/helper.rs:6:10
[INFO] [stdout]   |
[INFO] [stdout] 6 |         *try!(
[INFO] [stdout]   |          ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]  --> src/frameworks/cuda/helper.rs:7:13
[INFO] [stdout]   |
[INFO] [stdout] 7 |             try!(
[INFO] [stdout]   |             ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:167:44
[INFO] [stdout]     |
[INFO] [stdout] 167 |             ConvBackwardFilterAlgo::FFT => ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_FFT,
[INFO] [stdout]     |                                            ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 167 -             ConvBackwardFilterAlgo::FFT => ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_FFT,
[INFO] [stdout] 167 +             ConvBackwardFilterAlgo::FFT => cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_FFT,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]  --> src/frameworks/cuda/helper.rs:8:37
[INFO] [stdout]   |
[INFO] [stdout] 8 |                 x.get(device).ok_or(::co::plugin::Error::MissingMemoryForDevice("Unable to resolve memory."))
[INFO] [stdout]   |                                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]   |
[INFO] [stdout] 8 -                 x.get(device).ok_or(::co::plugin::Error::MissingMemoryForDevice("Unable to resolve memory."))
[INFO] [stdout] 8 +                 x.get(device).ok_or(co::plugin::Error::MissingMemoryForDevice("Unable to resolve memory."))
[INFO] [stdout]   |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:174:13
[INFO] [stdout]     |
[INFO] [stdout] 174 |             ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_0 => ConvBackwardFilterAlgo::ImplicitGEMMSum,
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 174 -             ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_0 => ConvBackwardFilterAlgo::ImplicitGEMMSum,
[INFO] [stdout] 174 +             cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_0 => ConvBackwardFilterAlgo::ImplicitGEMMSum,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]  --> src/frameworks/cuda/helper.rs:9:31
[INFO] [stdout]   |
[INFO] [stdout] 9 |             ).as_cuda().ok_or(::co::plugin::Error::MissingMemoryForDevice("Unable to receive CUDA memory."))
[INFO] [stdout]   |                               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]   |
[INFO] [stdout] 9 -             ).as_cuda().ok_or(::co::plugin::Error::MissingMemoryForDevice("Unable to receive CUDA memory."))
[INFO] [stdout] 9 +             ).as_cuda().ok_or(co::plugin::Error::MissingMemoryForDevice("Unable to receive CUDA memory."))
[INFO] [stdout]   |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:175:13
[INFO] [stdout]     |
[INFO] [stdout] 175 |             ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_1 => ConvBackwardFilterAlgo::ImplicitGEMM,
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 175 -             ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_1 => ConvBackwardFilterAlgo::ImplicitGEMM,
[INFO] [stdout] 175 +             cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_1 => ConvBackwardFilterAlgo::ImplicitGEMM,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]   --> src/frameworks/cuda/helper.rs:15:49
[INFO] [stdout]    |
[INFO] [stdout] 15 | pub unsafe fn receive_memory_ptr_mut<T>(x: &mut ::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*mut ::libc::...
[INFO] [stdout]    |                                                 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]    |
[INFO] [stdout] 15 - pub unsafe fn receive_memory_ptr_mut<T>(x: &mut ::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*mut ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout] 15 + pub unsafe fn receive_memory_ptr_mut<T>(x: &mut co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*mut ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout]    |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:176:13
[INFO] [stdout]     |
[INFO] [stdout] 176 |             ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_FFT => ConvBackwardFilterAlgo::FFT,
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 176 -             ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_FFT => ConvBackwardFilterAlgo::FFT,
[INFO] [stdout] 176 +             cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_FFT => ConvBackwardFilterAlgo::FFT,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]   --> src/frameworks/cuda/helper.rs:15:89
[INFO] [stdout]    |
[INFO] [stdout] 15 | pub unsafe fn receive_memory_ptr_mut<T>(x: &mut ::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*mut ::libc::...
[INFO] [stdout]    |                                                                                         ^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]    |
[INFO] [stdout] 15 - pub unsafe fn receive_memory_ptr_mut<T>(x: &mut ::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*mut ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout] 15 + pub unsafe fn receive_memory_ptr_mut<T>(x: &mut ::co::tensor::SharedTensor<T>, device: &co::device::DeviceType) -> Result<*mut ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout]    |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:177:13
[INFO] [stdout]     |
[INFO] [stdout] 177 | ...   ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_3 => ConvBackwardFilterAlgo::ImplicitPrecompiledGEMMSum,
[INFO] [stdout]     |       ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 177 -             ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_3 => ConvBackwardFilterAlgo::ImplicitPrecompiledGEMMSum,
[INFO] [stdout] 177 +             cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_3 => ConvBackwardFilterAlgo::ImplicitPrecompiledGEMMSum,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:188:41
[INFO] [stdout]     |
[INFO] [stdout] 188 |     ) -> Result<ConvBackwardFilterAlgo, ::co::error::Error> {
[INFO] [stdout]     |                                         ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 188 -     ) -> Result<ConvBackwardFilterAlgo, ::co::error::Error> {
[INFO] [stdout] 188 +     ) -> Result<ConvBackwardFilterAlgo, co::error::Error> {
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]   --> src/frameworks/cuda/helper.rs:15:130
[INFO] [stdout]    |
[INFO] [stdout] 15 | ...ensor<T>, device: &::co::device::DeviceType) -> Result<*mut ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout]    |                                                                ^^^^^^^^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]    |
[INFO] [stdout] 15 - pub unsafe fn receive_memory_ptr_mut<T>(x: &mut ::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*mut ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout] 15 + pub unsafe fn receive_memory_ptr_mut<T>(x: &mut ::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*mut libc::c_void, ::co::plugin::Error> {
[INFO] [stdout]    |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:194:29
[INFO] [stdout]     |
[INFO] [stdout] 194 | ...   0 => return Err(::co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution backward filter al...
[INFO] [stdout]     |                       ^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 194 -             0 => return Err(::co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution backward filter algorithm."))),
[INFO] [stdout] 194 +             0 => return Err(co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution backward filter algorithm."))),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]   --> src/frameworks/cuda/helper.rs:15:146
[INFO] [stdout]    |
[INFO] [stdout] 15 | ...::co::device::DeviceType) -> Result<*mut ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout]    |                                                             ^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]    |
[INFO] [stdout] 15 - pub unsafe fn receive_memory_ptr_mut<T>(x: &mut ::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*mut ::libc::c_void, ::co::plugin::Error> {
[INFO] [stdout] 15 + pub unsafe fn receive_memory_ptr_mut<T>(x: &mut ::co::tensor::SharedTensor<T>, device: &::co::device::DeviceType) -> Result<*mut ::libc::c_void, co::plugin::Error> {
[INFO] [stdout]    |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:194:56
[INFO] [stdout]     |
[INFO] [stdout] 194 | ...   0 => return Err(::co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution backward filter al...
[INFO] [stdout]     |                                                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 194 -             0 => return Err(::co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution backward filter algorithm."))),
[INFO] [stdout] 194 +             0 => return Err(::co::error::Error::Plugin(co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution backward filter algorithm."))),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]   --> src/frameworks/cuda/helper.rs:16:8
[INFO] [stdout]    |
[INFO] [stdout] 16 |     Ok(::std::mem::transmute::<u64, *mut ::libc::c_void>(
[INFO] [stdout]    |        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]    |
[INFO] [stdout] 16 -     Ok(::std::mem::transmute::<u64, *mut ::libc::c_void>(
[INFO] [stdout] 16 +     Ok(std::mem::transmute::<u64, *mut ::libc::c_void>(
[INFO] [stdout]    |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:203:65
[INFO] [stdout]     |
[INFO] [stdout] 203 |     fn as_cudnn(&self) -> Result<cudnnConvolutionBwdDataAlgo_t, ::co::error::Error> {
[INFO] [stdout]     |                                                                 ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 203 -     fn as_cudnn(&self) -> Result<cudnnConvolutionBwdDataAlgo_t, ::co::error::Error> {
[INFO] [stdout] 203 +     fn as_cudnn(&self) -> Result<cudnnConvolutionBwdDataAlgo_t, co::error::Error> {
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]   --> src/frameworks/cuda/helper.rs:16:42
[INFO] [stdout]    |
[INFO] [stdout] 16 |     Ok(::std::mem::transmute::<u64, *mut ::libc::c_void>(
[INFO] [stdout]    |                                          ^^^^^^^^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]    |
[INFO] [stdout] 16 -     Ok(::std::mem::transmute::<u64, *mut ::libc::c_void>(
[INFO] [stdout] 16 +     Ok(::std::mem::transmute::<u64, *mut libc::c_void>(
[INFO] [stdout]    |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]   --> src/frameworks/cuda/helper.rs:17:10
[INFO] [stdout]    |
[INFO] [stdout] 17 |         *try!(
[INFO] [stdout]    |          ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:205:54
[INFO] [stdout]     |
[INFO] [stdout] 205 | ...   ConvBackwardDataAlgo::Auto => return Err(::co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution bac...
[INFO] [stdout]     |                                                ^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 205 -             ConvBackwardDataAlgo::Auto => return Err(::co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution backward data algorithm from ConvBackwardDataAlgo::Auto. Use `find_cudnn_algo` to find an algorithm."))),
[INFO] [stdout] 205 +             ConvBackwardDataAlgo::Auto => return Err(co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution backward data algorithm from ConvBackwardDataAlgo::Auto. Use `find_cudnn_algo` to find an algorithm."))),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]   --> src/frameworks/cuda/helper.rs:18:13
[INFO] [stdout]    |
[INFO] [stdout] 18 |             try!(
[INFO] [stdout]    |             ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:205:81
[INFO] [stdout]     |
[INFO] [stdout] 205 | ...   ConvBackwardDataAlgo::Auto => return Err(::co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution bac...
[INFO] [stdout]     |                                                                           ^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 205 -             ConvBackwardDataAlgo::Auto => return Err(::co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution backward data algorithm from ConvBackwardDataAlgo::Auto. Use `find_cudnn_algo` to find an algorithm."))),
[INFO] [stdout] 205 +             ConvBackwardDataAlgo::Auto => return Err(::co::error::Error::Plugin(co::plugin::Error::Plugin("Can't create cuDNN convolution backward data algorithm from ConvBackwardDataAlgo::Auto. Use `find_cudnn_algo` to find an algorithm."))),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]   --> src/frameworks/cuda/helper.rs:19:41
[INFO] [stdout]    |
[INFO] [stdout] 19 |                 x.get_mut(device).ok_or(::co::plugin::Error::MissingMemoryForDevice("Unable to resolve memory."))
[INFO] [stdout]    |                                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]    |
[INFO] [stdout] 19 -                 x.get_mut(device).ok_or(::co::plugin::Error::MissingMemoryForDevice("Unable to resolve memory."))
[INFO] [stdout] 19 +                 x.get_mut(device).ok_or(co::plugin::Error::MissingMemoryForDevice("Unable to resolve memory."))
[INFO] [stdout]    |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]   --> src/frameworks/cuda/helper.rs:20:35
[INFO] [stdout]    |
[INFO] [stdout] 20 |             ).as_mut_cuda().ok_or(::co::plugin::Error::MissingMemoryForDevice("Unable to receive CUDA memory."))
[INFO] [stdout]    |                                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]    |
[INFO] [stdout] 20 -             ).as_mut_cuda().ok_or(::co::plugin::Error::MissingMemoryForDevice("Unable to receive CUDA memory."))
[INFO] [stdout] 20 +             ).as_mut_cuda().ok_or(co::plugin::Error::MissingMemoryForDevice("Unable to receive CUDA memory."))
[INFO] [stdout]    |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:206:51
[INFO] [stdout]     |
[INFO] [stdout] 206 |             ConvBackwardDataAlgo::ImplicitGEMM => ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_1,
[INFO] [stdout]     |                                                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 206 -             ConvBackwardDataAlgo::ImplicitGEMM => ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_1,
[INFO] [stdout] 206 +             ConvBackwardDataAlgo::ImplicitGEMM => cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_1,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:106:40
[INFO] [stdout]     |
[INFO] [stdout] 106 | impl_icudnndesc_for_sharedtensor!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     |                                        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 106 - impl_icudnndesc_for_sharedtensor!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout] 106 + impl_icudnndesc_for_sharedtensor!(f32, utils::DataType::Float);
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:207:54
[INFO] [stdout]     |
[INFO] [stdout] 207 |             ConvBackwardDataAlgo::ImplicitGEMMSum => ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_0,
[INFO] [stdout]     |                                                      ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 207 -             ConvBackwardDataAlgo::ImplicitGEMMSum => ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_0,
[INFO] [stdout] 207 +             ConvBackwardDataAlgo::ImplicitGEMMSum => cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_0,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:107:40
[INFO] [stdout]     |
[INFO] [stdout] 107 | impl_icudnndesc_for_sharedtensor!(f64, ::cudnn::utils::DataType::Double);
[INFO] [stdout]     |                                        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 107 - impl_icudnndesc_for_sharedtensor!(f64, ::cudnn::utils::DataType::Double);
[INFO] [stdout] 107 + impl_icudnndesc_for_sharedtensor!(f64, utils::DataType::Double);
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:208:42
[INFO] [stdout]     |
[INFO] [stdout] 208 |             ConvBackwardDataAlgo::FFT => ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT,
[INFO] [stdout]     |                                          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 208 -             ConvBackwardDataAlgo::FFT => ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT,
[INFO] [stdout] 208 +             ConvBackwardDataAlgo::FFT => cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:115:61
[INFO] [stdout]     |
[INFO] [stdout] 115 |     fn as_cudnn(&self) -> Result<cudnnConvolutionFwdAlgo_t, ::co::error::Error> {
[INFO] [stdout]     |                                                             ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 115 -     fn as_cudnn(&self) -> Result<cudnnConvolutionFwdAlgo_t, ::co::error::Error> {
[INFO] [stdout] 115 +     fn as_cudnn(&self) -> Result<cudnnConvolutionFwdAlgo_t, co::error::Error> {
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:209:48
[INFO] [stdout]     |
[INFO] [stdout] 209 |             ConvBackwardDataAlgo::FFTTiling => ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT_TILING,
[INFO] [stdout]     |                                                ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 209 -             ConvBackwardDataAlgo::FFTTiling => ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT_TILING,
[INFO] [stdout] 209 +             ConvBackwardDataAlgo::FFTTiling => cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT_TILING,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:117:49
[INFO] [stdout]     |
[INFO] [stdout] 117 | ...   ConvForwardAlgo::Auto => return Err(::co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution forward ...
[INFO] [stdout]     |                                           ^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 117 -             ConvForwardAlgo::Auto => return Err(::co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution forward algorithm from ConvForwardAlgo::Auto. Use `find_cudnn_algo` to find an algorithm."))),
[INFO] [stdout] 117 +             ConvForwardAlgo::Auto => return Err(co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution forward algorithm from ConvForwardAlgo::Auto. Use `find_cudnn_algo` to find an algorithm."))),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:216:13
[INFO] [stdout]     |
[INFO] [stdout] 216 |             ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_0 => ConvBackwardDataAlgo::ImplicitGEMMSum,
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 216 -             ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_0 => ConvBackwardDataAlgo::ImplicitGEMMSum,
[INFO] [stdout] 216 +             cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_0 => ConvBackwardDataAlgo::ImplicitGEMMSum,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:117:76
[INFO] [stdout]     |
[INFO] [stdout] 117 | ...   ConvForwardAlgo::Auto => return Err(::co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution forward ...
[INFO] [stdout]     |                                                                      ^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 117 -             ConvForwardAlgo::Auto => return Err(::co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution forward algorithm from ConvForwardAlgo::Auto. Use `find_cudnn_algo` to find an algorithm."))),
[INFO] [stdout] 117 +             ConvForwardAlgo::Auto => return Err(::co::error::Error::Plugin(co::plugin::Error::Plugin("Can't create cuDNN convolution forward algorithm from ConvForwardAlgo::Auto. Use `find_cudnn_algo` to find an algorithm."))),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:217:13
[INFO] [stdout]     |
[INFO] [stdout] 217 |             ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_1 => ConvBackwardDataAlgo::ImplicitGEMM,
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 217 -             ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_1 => ConvBackwardDataAlgo::ImplicitGEMM,
[INFO] [stdout] 217 +             cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_1 => ConvBackwardDataAlgo::ImplicitGEMM,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:118:38
[INFO] [stdout]     |
[INFO] [stdout] 118 |             ConvForwardAlgo::GEMM => ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_GEMM,
[INFO] [stdout]     |                                      ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 118 -             ConvForwardAlgo::GEMM => ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_GEMM,
[INFO] [stdout] 118 +             ConvForwardAlgo::GEMM => cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_GEMM,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:218:13
[INFO] [stdout]     |
[INFO] [stdout] 218 |             ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT => ConvBackwardDataAlgo::FFT,
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 218 -             ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT => ConvBackwardDataAlgo::FFT,
[INFO] [stdout] 218 +             cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT => ConvBackwardDataAlgo::FFT,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:119:46
[INFO] [stdout]     |
[INFO] [stdout] 119 |             ConvForwardAlgo::ImplicitGEMM => ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM,
[INFO] [stdout]     |                                              ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 119 -             ConvForwardAlgo::ImplicitGEMM => ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM,
[INFO] [stdout] 119 +             ConvForwardAlgo::ImplicitGEMM => cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:219:13
[INFO] [stdout]     |
[INFO] [stdout] 219 |             ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT_TILING => ConvBackwardDataAlgo::FFTTiling,
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 219 -             ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT_TILING => ConvBackwardDataAlgo::FFTTiling,
[INFO] [stdout] 219 +             cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT_TILING => ConvBackwardDataAlgo::FFTTiling,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:120:57
[INFO] [stdout]     |
[INFO] [stdout] 120 | ...   ConvForwardAlgo::ImplicitPrecompiledGEMM => ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_PRECOMP_GEMM,
[INFO] [stdout]     |                                                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 120 -             ConvForwardAlgo::ImplicitPrecompiledGEMM => ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_PRECOMP_GEMM,
[INFO] [stdout] 120 +             ConvForwardAlgo::ImplicitPrecompiledGEMM => cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_PRECOMP_GEMM,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:230:39
[INFO] [stdout]     |
[INFO] [stdout] 230 |     ) -> Result<ConvBackwardDataAlgo, ::co::error::Error> {
[INFO] [stdout]     |                                       ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 230 -     ) -> Result<ConvBackwardDataAlgo, ::co::error::Error> {
[INFO] [stdout] 230 +     ) -> Result<ConvBackwardDataAlgo, co::error::Error> {
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:121:37
[INFO] [stdout]     |
[INFO] [stdout] 121 |             ConvForwardAlgo::FFT => ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_FFT,
[INFO] [stdout]     |                                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 121 -             ConvForwardAlgo::FFT => ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_FFT,
[INFO] [stdout] 121 +             ConvForwardAlgo::FFT => cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_FFT,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:236:29
[INFO] [stdout]     |
[INFO] [stdout] 236 | ...   0 => return Err(::co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution backward data algo...
[INFO] [stdout]     |                       ^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 236 -             0 => return Err(::co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution backward data algorithm."))),
[INFO] [stdout] 236 +             0 => return Err(co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution backward data algorithm."))),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:122:43
[INFO] [stdout]     |
[INFO] [stdout] 122 |             ConvForwardAlgo::FFTTiling => ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_FFT_TILING,
[INFO] [stdout]     |                                           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 122 -             ConvForwardAlgo::FFTTiling => ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_FFT_TILING,
[INFO] [stdout] 122 +             ConvForwardAlgo::FFTTiling => cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_FFT_TILING,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:236:56
[INFO] [stdout]     |
[INFO] [stdout] 236 | ...   0 => return Err(::co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution backward data algo...
[INFO] [stdout]     |                                                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 236 -             0 => return Err(::co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution backward data algorithm."))),
[INFO] [stdout] 236 +             0 => return Err(::co::error::Error::Plugin(co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution backward data algorithm."))),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:123:40
[INFO] [stdout]     |
[INFO] [stdout] 123 |             ConvForwardAlgo::Direct => ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_DIRECT,
[INFO] [stdout]     |                                        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 123 -             ConvForwardAlgo::Direct => ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_DIRECT,
[INFO] [stdout] 123 +             ConvForwardAlgo::Direct => cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_DIRECT,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:263:32
[INFO] [stdout]     |
[INFO] [stdout] 263 |                 let src_desc = try!(src.cudnn_tensor_desc());
[INFO] [stdout]     |                                ^^^
[INFO] [stdout] ...
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:264:33
[INFO] [stdout]     |
[INFO] [stdout] 264 |                 let dest_desc = try!(dest.cudnn_tensor_desc());
[INFO] [stdout]     |                                 ^^^
[INFO] [stdout] ...
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:130:13
[INFO] [stdout]     |
[INFO] [stdout] 130 |             ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_GEMM => ConvForwardAlgo::GEMM,
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 130 -             ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_GEMM => ConvForwardAlgo::GEMM,
[INFO] [stdout] 130 +             cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_GEMM => ConvForwardAlgo::GEMM,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:131:13
[INFO] [stdout]     |
[INFO] [stdout] 131 |             ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM => ConvForwardAlgo::ImplicitGEMM,
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 131 -             ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM => ConvForwardAlgo::ImplicitGEMM,
[INFO] [stdout] 131 +             cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_GEMM => ConvForwardAlgo::ImplicitGEMM,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:265:35
[INFO] [stdout]     |
[INFO] [stdout] 265 |                 let filter_desc = try!(filter.cudnn_filter_desc());
[INFO] [stdout]     |                                   ^^^
[INFO] [stdout] ...
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:132:13
[INFO] [stdout]     |
[INFO] [stdout] 132 | ...   ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_PRECOMP_GEMM => ConvForwardAlgo::ImplicitPrecompiledGEMM,
[INFO] [stdout]     |       ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 132 -             ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_PRECOMP_GEMM => ConvForwardAlgo::ImplicitPrecompiledGEMM,
[INFO] [stdout] 132 +             cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_IMPLICIT_PRECOMP_GEMM => ConvForwardAlgo::ImplicitPrecompiledGEMM,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:133:13
[INFO] [stdout]     |
[INFO] [stdout] 133 |             ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_FFT => ConvForwardAlgo::FFT,
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 133 -             ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_FFT => ConvForwardAlgo::FFT,
[INFO] [stdout] 133 +             cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_FFT => ConvForwardAlgo::FFT,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:310:41
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     |                                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 310 - impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout] 310 + impl_convolution_for_cuda_backend!(f32, utils::DataType::Float);
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:134:13
[INFO] [stdout]     |
[INFO] [stdout] 134 |             ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_FFT_TILING => ConvForwardAlgo::FFTTiling,
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 134 -             ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_FFT_TILING => ConvForwardAlgo::FFTTiling,
[INFO] [stdout] 134 +             cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_FFT_TILING => ConvForwardAlgo::FFTTiling,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:268:40
[INFO] [stdout]     |
[INFO] [stdout] 268 |                 let useable_algo_fwd = try!(algo_fwd.find_cudnn_algo(&filter_desc, &conv_desc, &src_desc, &dest_desc));
[INFO] [stdout]     |                                        ^^^
[INFO] [stdout] ...
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:135:13
[INFO] [stdout]     |
[INFO] [stdout] 135 |             ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_DIRECT => ConvForwardAlgo::Direct,
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 135 -             ::cudnn::cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_DIRECT => ConvForwardAlgo::Direct,
[INFO] [stdout] 135 +             cudnnConvolutionFwdAlgo_t::CUDNN_CONVOLUTION_FWD_ALGO_DIRECT => ConvForwardAlgo::Direct,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:269:47
[INFO] [stdout]     |
[INFO] [stdout] 269 |                 let useable_algo_bwd_filter = try!(algo_bwd_filter.find_cudnn_algo(&filter_desc, &conv_desc, &src_desc, &dest_desc));
[INFO] [stdout]     |                                               ^^^
[INFO] [stdout] ...
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:146:34
[INFO] [stdout]     |
[INFO] [stdout] 146 |     ) -> Result<ConvForwardAlgo, ::co::error::Error> {
[INFO] [stdout]     |                                  ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 146 -     ) -> Result<ConvForwardAlgo, ::co::error::Error> {
[INFO] [stdout] 146 +     ) -> Result<ConvForwardAlgo, co::error::Error> {
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:270:45
[INFO] [stdout]     |
[INFO] [stdout] 270 |                 let useable_algo_bwd_data = try!(algo_bwd_data.find_cudnn_algo(&filter_desc, &conv_desc, &src_desc, &dest_desc));
[INFO] [stdout]     |                                             ^^^
[INFO] [stdout] ...
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:152:29
[INFO] [stdout]     |
[INFO] [stdout] 152 | ...   0 => return Err(::co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution forward algorithm....
[INFO] [stdout]     |                       ^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 152 -             0 => return Err(::co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution forward algorithm."))),
[INFO] [stdout] 152 +             0 => return Err(co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution forward algorithm."))),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:152:56
[INFO] [stdout]     |
[INFO] [stdout] 152 | ...   0 => return Err(::co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution forward algorithm....
[INFO] [stdout]     |                                                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 152 -             0 => return Err(::co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution forward algorithm."))),
[INFO] [stdout] 152 +             0 => return Err(::co::error::Error::Plugin(co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution forward algorithm."))),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:516:54
[INFO] [stdout]     |
[INFO] [stdout] 516 |             match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:161:67
[INFO] [stdout]     |
[INFO] [stdout] 161 |     fn as_cudnn(&self) -> Result<cudnnConvolutionBwdFilterAlgo_t, ::co::error::Error> {
[INFO] [stdout]     |                                                                   ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 161 -     fn as_cudnn(&self) -> Result<cudnnConvolutionBwdFilterAlgo_t, ::co::error::Error> {
[INFO] [stdout] 161 +     fn as_cudnn(&self) -> Result<cudnnConvolutionBwdFilterAlgo_t, co::error::Error> {
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:533:16
[INFO] [stdout]     |
[INFO] [stdout] 533 |             Ok(try!(match CUDNN.convolution_forward(
[INFO] [stdout]     |                ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:163:56
[INFO] [stdout]     |
[INFO] [stdout] 163 | ...   ConvBackwardFilterAlgo::Auto => return Err(::co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution b...
[INFO] [stdout]     |                                                  ^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 163 -             ConvBackwardFilterAlgo::Auto => return Err(::co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution backward filter algorithm from ConvBackwardFilterAlgo::Auto. Use `find_cudnn_algo` to find an algorithm."))),
[INFO] [stdout] 163 +             ConvBackwardFilterAlgo::Auto => return Err(co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution backward filter algorithm from ConvBackwardFilterAlgo::Auto. Use `find_cudnn_algo` to find an algorithm."))),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:535:17
[INFO] [stdout]     |
[INFO] [stdout] 535 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(workspace, self.device()) }),
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:163:83
[INFO] [stdout]     |
[INFO] [stdout] 163 | ...   ConvBackwardFilterAlgo::Auto => return Err(::co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution b...
[INFO] [stdout]     |                                                                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 163 -             ConvBackwardFilterAlgo::Auto => return Err(::co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution backward filter algorithm from ConvBackwardFilterAlgo::Auto. Use `find_cudnn_algo` to find an algorithm."))),
[INFO] [stdout] 163 +             ConvBackwardFilterAlgo::Auto => return Err(::co::error::Error::Plugin(co::plugin::Error::Plugin("Can't create cuDNN convolution backward filter algorithm from ConvBackwardFilterAlgo::Auto. Use `find_cudnn_algo` to find an algorithm."))),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:164:53
[INFO] [stdout]     |
[INFO] [stdout] 164 |             ConvBackwardFilterAlgo::ImplicitGEMM => ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_1,
[INFO] [stdout]     |                                                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 164 -             ConvBackwardFilterAlgo::ImplicitGEMM => ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_1,
[INFO] [stdout] 164 +             ConvBackwardFilterAlgo::ImplicitGEMM => cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_1,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:536:17
[INFO] [stdout]     |
[INFO] [stdout] 536 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(filter, self.device()) }),
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:165:56
[INFO] [stdout]     |
[INFO] [stdout] 165 |             ConvBackwardFilterAlgo::ImplicitGEMMSum => ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_0,
[INFO] [stdout]     |                                                        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 165 -             ConvBackwardFilterAlgo::ImplicitGEMMSum => ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_0,
[INFO] [stdout] 165 +             ConvBackwardFilterAlgo::ImplicitGEMMSum => cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_0,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:538:17
[INFO] [stdout]     |
[INFO] [stdout] 538 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:166:67
[INFO] [stdout]     |
[INFO] [stdout] 166 | ...plicitPrecompiledGEMMSum => ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_3,
[INFO] [stdout]     |                                ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 166 -             ConvBackwardFilterAlgo::ImplicitPrecompiledGEMMSum => ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_3,
[INFO] [stdout] 166 +             ConvBackwardFilterAlgo::ImplicitPrecompiledGEMMSum => cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_3,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:167:44
[INFO] [stdout]     |
[INFO] [stdout] 167 |             ConvBackwardFilterAlgo::FFT => ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_FFT,
[INFO] [stdout]     |                                            ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 167 -             ConvBackwardFilterAlgo::FFT => ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_FFT,
[INFO] [stdout] 167 +             ConvBackwardFilterAlgo::FFT => cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_FFT,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:540:17
[INFO] [stdout]     |
[INFO] [stdout] 540 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result, self.device()) }), // dest_data
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:174:13
[INFO] [stdout]     |
[INFO] [stdout] 174 |             ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_0 => ConvBackwardFilterAlgo::ImplicitGEMMSum,
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 174 -             ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_0 => ConvBackwardFilterAlgo::ImplicitGEMMSum,
[INFO] [stdout] 174 +             cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_0 => ConvBackwardFilterAlgo::ImplicitGEMMSum,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:537:18
[INFO] [stdout]     |
[INFO] [stdout] 537 |                 &try!(x.cudnn_tensor_desc()), // src_desc
[INFO] [stdout]     |                  ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:175:13
[INFO] [stdout]     |
[INFO] [stdout] 175 |             ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_1 => ConvBackwardFilterAlgo::ImplicitGEMM,
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 175 -             ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_1 => ConvBackwardFilterAlgo::ImplicitGEMM,
[INFO] [stdout] 175 +             cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_1 => ConvBackwardFilterAlgo::ImplicitGEMM,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:539:18
[INFO] [stdout]     |
[INFO] [stdout] 539 |                 &try!(result.cudnn_tensor_desc()), // dest_desc
[INFO] [stdout]     |                  ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:176:13
[INFO] [stdout]     |
[INFO] [stdout] 176 |             ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_FFT => ConvBackwardFilterAlgo::FFT,
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 176 -             ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_FFT => ConvBackwardFilterAlgo::FFT,
[INFO] [stdout] 176 +             cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_FFT => ConvBackwardFilterAlgo::FFT,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:559:61
[INFO] [stdout]     |
[INFO] [stdout] 559 |             match src_data.add_device(self.device()) { _ => try!(src_data.sync(self.device())) }
[INFO] [stdout]     |                                                             ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:177:13
[INFO] [stdout]     |
[INFO] [stdout] 177 | ...   ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_3 => ConvBackwardFilterAlgo::ImplicitPrecompiledGEMMSum,
[INFO] [stdout]     |       ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 177 -             ::cudnn::cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_3 => ConvBackwardFilterAlgo::ImplicitPrecompiledGEMMSum,
[INFO] [stdout] 177 +             cudnnConvolutionBwdFilterAlgo_t::CUDNN_CONVOLUTION_BWD_FILTER_ALGO_3 => ConvBackwardFilterAlgo::ImplicitPrecompiledGEMMSum,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:560:62
[INFO] [stdout]     |
[INFO] [stdout] 560 |             match dest_diff.add_device(self.device()) { _ => try!(dest_diff.sync(self.device())) }
[INFO] [stdout]     |                                                              ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:188:41
[INFO] [stdout]     |
[INFO] [stdout] 188 |     ) -> Result<ConvBackwardFilterAlgo, ::co::error::Error> {
[INFO] [stdout]     |                                         ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 188 -     ) -> Result<ConvBackwardFilterAlgo, ::co::error::Error> {
[INFO] [stdout] 188 +     ) -> Result<ConvBackwardFilterAlgo, co::error::Error> {
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:194:29
[INFO] [stdout]     |
[INFO] [stdout] 194 | ...   0 => return Err(::co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution backward filter al...
[INFO] [stdout]     |                       ^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 194 -             0 => return Err(::co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution backward filter algorithm."))),
[INFO] [stdout] 194 +             0 => return Err(co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution backward filter algorithm."))),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:561:64
[INFO] [stdout]     |
[INFO] [stdout] 561 |             match filter_diff.add_device(self.device()) { _ => try!(filter_diff.sync(self.device())) }
[INFO] [stdout]     |                                                                ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:194:56
[INFO] [stdout]     |
[INFO] [stdout] 194 | ...   0 => return Err(::co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution backward filter al...
[INFO] [stdout]     |                                                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 194 -             0 => return Err(::co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution backward filter algorithm."))),
[INFO] [stdout] 194 +             0 => return Err(::co::error::Error::Plugin(co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution backward filter algorithm."))),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:203:65
[INFO] [stdout]     |
[INFO] [stdout] 203 |     fn as_cudnn(&self) -> Result<cudnnConvolutionBwdDataAlgo_t, ::co::error::Error> {
[INFO] [stdout]     |                                                                 ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 203 -     fn as_cudnn(&self) -> Result<cudnnConvolutionBwdDataAlgo_t, ::co::error::Error> {
[INFO] [stdout] 203 +     fn as_cudnn(&self) -> Result<cudnnConvolutionBwdDataAlgo_t, co::error::Error> {
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:578:16
[INFO] [stdout]     |
[INFO] [stdout] 578 |             Ok(try!(match CUDNN.convolution_backward_filter(
[INFO] [stdout]     |                ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:205:54
[INFO] [stdout]     |
[INFO] [stdout] 205 | ...   ConvBackwardDataAlgo::Auto => return Err(::co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution bac...
[INFO] [stdout]     |                                                ^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 205 -             ConvBackwardDataAlgo::Auto => return Err(::co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution backward data algorithm from ConvBackwardDataAlgo::Auto. Use `find_cudnn_algo` to find an algorithm."))),
[INFO] [stdout] 205 +             ConvBackwardDataAlgo::Auto => return Err(co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution backward data algorithm from ConvBackwardDataAlgo::Auto. Use `find_cudnn_algo` to find an algorithm."))),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:580:17
[INFO] [stdout]     |
[INFO] [stdout] 580 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(workspace, self.device()) }),
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:205:81
[INFO] [stdout]     |
[INFO] [stdout] 205 | ...   ConvBackwardDataAlgo::Auto => return Err(::co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution bac...
[INFO] [stdout]     |                                                                           ^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 205 -             ConvBackwardDataAlgo::Auto => return Err(::co::error::Error::Plugin(::co::plugin::Error::Plugin("Can't create cuDNN convolution backward data algorithm from ConvBackwardDataAlgo::Auto. Use `find_cudnn_algo` to find an algorithm."))),
[INFO] [stdout] 205 +             ConvBackwardDataAlgo::Auto => return Err(::co::error::Error::Plugin(co::plugin::Error::Plugin("Can't create cuDNN convolution backward data algorithm from ConvBackwardDataAlgo::Auto. Use `find_cudnn_algo` to find an algorithm."))),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:582:17
[INFO] [stdout]     |
[INFO] [stdout] 582 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(src_data, self.device()) }),
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:206:51
[INFO] [stdout]     |
[INFO] [stdout] 206 |             ConvBackwardDataAlgo::ImplicitGEMM => ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_1,
[INFO] [stdout]     |                                                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 206 -             ConvBackwardDataAlgo::ImplicitGEMM => ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_1,
[INFO] [stdout] 206 +             ConvBackwardDataAlgo::ImplicitGEMM => cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_1,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:207:54
[INFO] [stdout]     |
[INFO] [stdout] 207 |             ConvBackwardDataAlgo::ImplicitGEMMSum => ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_0,
[INFO] [stdout]     |                                                      ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 207 -             ConvBackwardDataAlgo::ImplicitGEMMSum => ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_0,
[INFO] [stdout] 207 +             ConvBackwardDataAlgo::ImplicitGEMMSum => cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_0,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:584:17
[INFO] [stdout]     |
[INFO] [stdout] 584 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(dest_diff, self.device()) }),
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:208:42
[INFO] [stdout]     |
[INFO] [stdout] 208 |             ConvBackwardDataAlgo::FFT => ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT,
[INFO] [stdout]     |                                          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 208 -             ConvBackwardDataAlgo::FFT => ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT,
[INFO] [stdout] 208 +             ConvBackwardDataAlgo::FFT => cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:585:17
[INFO] [stdout]     |
[INFO] [stdout] 585 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(filter_diff, self.device()) }),
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:209:48
[INFO] [stdout]     |
[INFO] [stdout] 209 |             ConvBackwardDataAlgo::FFTTiling => ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT_TILING,
[INFO] [stdout]     |                                                ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 209 -             ConvBackwardDataAlgo::FFTTiling => ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT_TILING,
[INFO] [stdout] 209 +             ConvBackwardDataAlgo::FFTTiling => cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT_TILING,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:216:13
[INFO] [stdout]     |
[INFO] [stdout] 216 |             ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_0 => ConvBackwardDataAlgo::ImplicitGEMMSum,
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 216 -             ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_0 => ConvBackwardDataAlgo::ImplicitGEMMSum,
[INFO] [stdout] 216 +             cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_0 => ConvBackwardDataAlgo::ImplicitGEMMSum,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:581:18
[INFO] [stdout]     |
[INFO] [stdout] 581 |                 &try!(src_data.cudnn_tensor_desc()),
[INFO] [stdout]     |                  ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:217:13
[INFO] [stdout]     |
[INFO] [stdout] 217 |             ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_1 => ConvBackwardDataAlgo::ImplicitGEMM,
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 217 -             ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_1 => ConvBackwardDataAlgo::ImplicitGEMM,
[INFO] [stdout] 217 +             cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_1 => ConvBackwardDataAlgo::ImplicitGEMM,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:583:18
[INFO] [stdout]     |
[INFO] [stdout] 583 |                 &try!(dest_diff.cudnn_tensor_desc()),
[INFO] [stdout]     |                  ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:218:13
[INFO] [stdout]     |
[INFO] [stdout] 218 |             ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT => ConvBackwardDataAlgo::FFT,
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 218 -             ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT => ConvBackwardDataAlgo::FFT,
[INFO] [stdout] 218 +             cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT => ConvBackwardDataAlgo::FFT,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:219:13
[INFO] [stdout]     |
[INFO] [stdout] 219 |             ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT_TILING => ConvBackwardDataAlgo::FFTTiling,
[INFO] [stdout]     |             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 219 -             ::cudnn::cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT_TILING => ConvBackwardDataAlgo::FFTTiling,
[INFO] [stdout] 219 +             cudnnConvolutionBwdDataAlgo_t::CUDNN_CONVOLUTION_BWD_DATA_ALGO_FFT_TILING => ConvBackwardDataAlgo::FFTTiling,
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:604:59
[INFO] [stdout]     |
[INFO] [stdout] 604 |             match filter.add_device(self.device()) { _ => try!(filter.sync(self.device())) }
[INFO] [stdout]     |                                                           ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:605:59
[INFO] [stdout]     |
[INFO] [stdout] 605 |             match x_diff.add_device(self.device()) { _ => try!(x_diff.sync(self.device())) }
[INFO] [stdout]     |                                                           ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:230:39
[INFO] [stdout]     |
[INFO] [stdout] 230 |     ) -> Result<ConvBackwardDataAlgo, ::co::error::Error> {
[INFO] [stdout]     |                                       ^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 230 -     ) -> Result<ConvBackwardDataAlgo, ::co::error::Error> {
[INFO] [stdout] 230 +     ) -> Result<ConvBackwardDataAlgo, co::error::Error> {
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:236:29
[INFO] [stdout]     |
[INFO] [stdout] 236 | ...   0 => return Err(::co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution backward data algo...
[INFO] [stdout]     |                       ^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 236 -             0 => return Err(::co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution backward data algorithm."))),
[INFO] [stdout] 236 +             0 => return Err(co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution backward data algorithm."))),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:606:64
[INFO] [stdout]     |
[INFO] [stdout] 606 |             match result_diff.add_device(self.device()) { _ => try!(result_diff.sync(self.device())) }
[INFO] [stdout]     |                                                                ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:236:56
[INFO] [stdout]     |
[INFO] [stdout] 236 | ...   0 => return Err(::co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution backward data algo...
[INFO] [stdout]     |                                                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 236 -             0 => return Err(::co::error::Error::Plugin(::co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution backward data algorithm."))),
[INFO] [stdout] 236 +             0 => return Err(::co::error::Error::Plugin(co::plugin::Error::Operation("Unable to find CUDA cuDNN convolution backward data algorithm."))),
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:623:16
[INFO] [stdout]     |
[INFO] [stdout] 623 |             Ok(try!(match CUDNN.convolution_backward_data(
[INFO] [stdout]     |                ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:263:32
[INFO] [stdout]     |
[INFO] [stdout] 263 |                 let src_desc = try!(src.cudnn_tensor_desc());
[INFO] [stdout]     |                                ^^^
[INFO] [stdout] ...
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:625:17
[INFO] [stdout]     |
[INFO] [stdout] 625 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(workspace, self.device()) }),
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:264:33
[INFO] [stdout]     |
[INFO] [stdout] 264 |                 let dest_desc = try!(dest.cudnn_tensor_desc());
[INFO] [stdout]     |                                 ^^^
[INFO] [stdout] ...
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:626:17
[INFO] [stdout]     |
[INFO] [stdout] 626 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(filter, self.device()) }),
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:265:35
[INFO] [stdout]     |
[INFO] [stdout] 265 |                 let filter_desc = try!(filter.cudnn_filter_desc());
[INFO] [stdout]     |                                   ^^^
[INFO] [stdout] ...
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:310:41
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     |                                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 310 - impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout] 310 + impl_convolution_for_cuda_backend!(f32, utils::DataType::Float);
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:628:17
[INFO] [stdout]     |
[INFO] [stdout] 628 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x_diff, self.device()) }),
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:268:40
[INFO] [stdout]     |
[INFO] [stdout] 268 |                 let useable_algo_fwd = try!(algo_fwd.find_cudnn_algo(&filter_desc, &conv_desc, &src_desc, &dest_desc));
[INFO] [stdout]     |                                        ^^^
[INFO] [stdout] ...
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:269:47
[INFO] [stdout]     |
[INFO] [stdout] 269 |                 let useable_algo_bwd_filter = try!(algo_bwd_filter.find_cudnn_algo(&filter_desc, &conv_desc, &src_desc, &dest_desc));
[INFO] [stdout]     |                                               ^^^
[INFO] [stdout] ...
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:630:17
[INFO] [stdout]     |
[INFO] [stdout] 630 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result_diff, self.device()) }),
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:270:45
[INFO] [stdout]     |
[INFO] [stdout] 270 |                 let useable_algo_bwd_data = try!(algo_bwd_data.find_cudnn_algo(&filter_desc, &conv_desc, &src_desc, &dest_desc));
[INFO] [stdout]     |                                             ^^^
[INFO] [stdout] ...
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:627:18
[INFO] [stdout]     |
[INFO] [stdout] 627 |                 &try!(x_diff.cudnn_tensor_desc()),
[INFO] [stdout]     |                  ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:516:54
[INFO] [stdout]     |
[INFO] [stdout] 516 |             match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:629:18
[INFO] [stdout]     |
[INFO] [stdout] 629 |                 &try!(result_diff.cudnn_tensor_desc()),
[INFO] [stdout]     |                  ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:55:58
[INFO] [stdout]     |
[INFO] [stdout] 55  |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:533:16
[INFO] [stdout]     |
[INFO] [stdout] 533 |             Ok(try!(match CUDNN.convolution_forward(
[INFO] [stdout]     |                ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:68:20
[INFO] [stdout]     |
[INFO] [stdout] 68  |                 Ok(try!(match CUDNN.sigmoid_forward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:535:17
[INFO] [stdout]     |
[INFO] [stdout] 535 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(workspace, self.device()) }),
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:70:21
[INFO] [stdout]     |
[INFO] [stdout] 70  |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:536:17
[INFO] [stdout]     |
[INFO] [stdout] 536 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(filter, self.device()) }),
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:72:21
[INFO] [stdout]     |
[INFO] [stdout] 72  |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:69:22
[INFO] [stdout]     |
[INFO] [stdout] 69  |                     &try!(x.cudnn_tensor_desc_flat()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:538:17
[INFO] [stdout]     |
[INFO] [stdout] 538 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:71:22
[INFO] [stdout]     |
[INFO] [stdout] 71  |                     &try!(result.cudnn_tensor_desc_flat()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:540:17
[INFO] [stdout]     |
[INFO] [stdout] 540 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result, self.device()) }), // dest_data
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:89:58
[INFO] [stdout]     |
[INFO] [stdout] 89  |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:90:63
[INFO] [stdout]     |
[INFO] [stdout] 90  |                 match x_diff.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:537:18
[INFO] [stdout]     |
[INFO] [stdout] 537 |                 &try!(x.cudnn_tensor_desc()), // src_desc
[INFO] [stdout]     |                  ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:91:63
[INFO] [stdout]     |
[INFO] [stdout] 91  |                 match result.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:106:20
[INFO] [stdout]     |
[INFO] [stdout] 106 |                 Ok(try!(match CUDNN.sigmoid_backward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:539:18
[INFO] [stdout]     |
[INFO] [stdout] 539 |                 &try!(result.cudnn_tensor_desc()), // dest_desc
[INFO] [stdout]     |                  ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:108:21
[INFO] [stdout]     |
[INFO] [stdout] 108 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:559:61
[INFO] [stdout]     |
[INFO] [stdout] 559 |             match src_data.add_device(self.device()) { _ => try!(src_data.sync(self.device())) }
[INFO] [stdout]     |                                                             ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:110:21
[INFO] [stdout]     |
[INFO] [stdout] 110 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x_diff, self.device()) }), //src_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:112:21
[INFO] [stdout]     |
[INFO] [stdout] 112 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(result, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:560:62
[INFO] [stdout]     |
[INFO] [stdout] 560 |             match dest_diff.add_device(self.device()) { _ => try!(dest_diff.sync(self.device())) }
[INFO] [stdout]     |                                                              ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:114:21
[INFO] [stdout]     |
[INFO] [stdout] 114 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result_diff, self.device()) }), // dest_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:561:64
[INFO] [stdout]     |
[INFO] [stdout] 561 |             match filter_diff.add_device(self.device()) { _ => try!(filter_diff.sync(self.device())) }
[INFO] [stdout]     |                                                                ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:578:16
[INFO] [stdout]     |
[INFO] [stdout] 578 |             Ok(try!(match CUDNN.convolution_backward_filter(
[INFO] [stdout]     |                ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:107:22
[INFO] [stdout]     |
[INFO] [stdout] 107 |                     &try!(x.cudnn_tensor_desc_flat()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:109:22
[INFO] [stdout]     |
[INFO] [stdout] 109 |                     &try!(x_diff.cudnn_tensor_desc_flat()), // src_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:580:17
[INFO] [stdout]     |
[INFO] [stdout] 580 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(workspace, self.device()) }),
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:111:22
[INFO] [stdout]     |
[INFO] [stdout] 111 |                     &try!(result.cudnn_tensor_desc_flat()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:582:17
[INFO] [stdout]     |
[INFO] [stdout] 582 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(src_data, self.device()) }),
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:113:22
[INFO] [stdout]     |
[INFO] [stdout] 113 |                     &try!(result_diff.cudnn_tensor_desc_flat()), // dest_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:584:17
[INFO] [stdout]     |
[INFO] [stdout] 584 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(dest_diff, self.device()) }),
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:208:58
[INFO] [stdout]     |
[INFO] [stdout] 208 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:585:17
[INFO] [stdout]     |
[INFO] [stdout] 585 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(filter_diff, self.device()) }),
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:221:20
[INFO] [stdout]     |
[INFO] [stdout] 221 |                 Ok(try!(match CUDNN.relu_forward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:223:21
[INFO] [stdout]     |
[INFO] [stdout] 223 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:581:18
[INFO] [stdout]     |
[INFO] [stdout] 581 |                 &try!(src_data.cudnn_tensor_desc()),
[INFO] [stdout]     |                  ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:225:21
[INFO] [stdout]     |
[INFO] [stdout] 225 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:583:18
[INFO] [stdout]     |
[INFO] [stdout] 583 |                 &try!(dest_diff.cudnn_tensor_desc()),
[INFO] [stdout]     |                  ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:222:22
[INFO] [stdout]     |
[INFO] [stdout] 222 |                     &try!(x.cudnn_tensor_desc_flat()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:604:59
[INFO] [stdout]     |
[INFO] [stdout] 604 |             match filter.add_device(self.device()) { _ => try!(filter.sync(self.device())) }
[INFO] [stdout]     |                                                           ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:224:22
[INFO] [stdout]     |
[INFO] [stdout] 224 |                     &try!(result.cudnn_tensor_desc_flat()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:242:58
[INFO] [stdout]     |
[INFO] [stdout] 242 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:605:59
[INFO] [stdout]     |
[INFO] [stdout] 605 |             match x_diff.add_device(self.device()) { _ => try!(x_diff.sync(self.device())) }
[INFO] [stdout]     |                                                           ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:243:63
[INFO] [stdout]     |
[INFO] [stdout] 243 |                 match x_diff.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:244:63
[INFO] [stdout]     |
[INFO] [stdout] 244 |                 match result.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:606:64
[INFO] [stdout]     |
[INFO] [stdout] 606 |             match result_diff.add_device(self.device()) { _ => try!(result_diff.sync(self.device())) }
[INFO] [stdout]     |                                                                ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:259:20
[INFO] [stdout]     |
[INFO] [stdout] 259 |                 Ok(try!(match CUDNN.relu_backward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:623:16
[INFO] [stdout]     |
[INFO] [stdout] 623 |             Ok(try!(match CUDNN.convolution_backward_data(
[INFO] [stdout]     |                ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:261:21
[INFO] [stdout]     |
[INFO] [stdout] 261 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:263:21
[INFO] [stdout]     |
[INFO] [stdout] 263 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x_diff, self.device()) }), //src_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:625:17
[INFO] [stdout]     |
[INFO] [stdout] 625 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(workspace, self.device()) }),
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:265:21
[INFO] [stdout]     |
[INFO] [stdout] 265 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(result, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:626:17
[INFO] [stdout]     |
[INFO] [stdout] 626 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(filter, self.device()) }),
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:267:21
[INFO] [stdout]     |
[INFO] [stdout] 267 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result_diff, self.device()) }), // dest_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:628:17
[INFO] [stdout]     |
[INFO] [stdout] 628 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x_diff, self.device()) }),
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:260:22
[INFO] [stdout]     |
[INFO] [stdout] 260 |                     &try!(x.cudnn_tensor_desc_flat()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:630:17
[INFO] [stdout]     |
[INFO] [stdout] 630 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result_diff, self.device()) }),
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:262:22
[INFO] [stdout]     |
[INFO] [stdout] 262 |                     &try!(x_diff.cudnn_tensor_desc_flat()), // src_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:264:22
[INFO] [stdout]     |
[INFO] [stdout] 264 |                     &try!(result.cudnn_tensor_desc_flat()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:627:18
[INFO] [stdout]     |
[INFO] [stdout] 627 |                 &try!(x_diff.cudnn_tensor_desc()),
[INFO] [stdout]     |                  ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:266:22
[INFO] [stdout]     |
[INFO] [stdout] 266 |                     &try!(result_diff.cudnn_tensor_desc_flat()), // dest_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:629:18
[INFO] [stdout]     |
[INFO] [stdout] 629 |                 &try!(result_diff.cudnn_tensor_desc()),
[INFO] [stdout]     |                  ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:310:1
[INFO] [stdout]     |
[INFO] [stdout] 310 | impl_convolution_for_cuda_backend!(f32, ::cudnn::utils::DataType::Float);
[INFO] [stdout]     | ------------------------------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:361:58
[INFO] [stdout]     |
[INFO] [stdout] 361 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:55:58
[INFO] [stdout]     |
[INFO] [stdout] 55  |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:374:20
[INFO] [stdout]     |
[INFO] [stdout] 374 |                 Ok(try!(match CUDNN.tanh_forward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:68:20
[INFO] [stdout]     |
[INFO] [stdout] 68  |                 Ok(try!(match CUDNN.sigmoid_forward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:376:21
[INFO] [stdout]     |
[INFO] [stdout] 376 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:70:21
[INFO] [stdout]     |
[INFO] [stdout] 70  |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:378:21
[INFO] [stdout]     |
[INFO] [stdout] 378 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:72:21
[INFO] [stdout]     |
[INFO] [stdout] 72  |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:375:22
[INFO] [stdout]     |
[INFO] [stdout] 375 |                     &try!(x.cudnn_tensor_desc_flat()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:69:22
[INFO] [stdout]     |
[INFO] [stdout] 69  |                     &try!(x.cudnn_tensor_desc_flat()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:377:22
[INFO] [stdout]     |
[INFO] [stdout] 377 |                     &try!(result.cudnn_tensor_desc_flat()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:71:22
[INFO] [stdout]     |
[INFO] [stdout] 71  |                     &try!(result.cudnn_tensor_desc_flat()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:395:58
[INFO] [stdout]     |
[INFO] [stdout] 395 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:89:58
[INFO] [stdout]     |
[INFO] [stdout] 89  |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:396:63
[INFO] [stdout]     |
[INFO] [stdout] 396 |                 match x_diff.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:397:63
[INFO] [stdout]     |
[INFO] [stdout] 397 |                 match result.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:90:63
[INFO] [stdout]     |
[INFO] [stdout] 90  |                 match x_diff.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:412:20
[INFO] [stdout]     |
[INFO] [stdout] 412 |                 Ok(try!(match CUDNN.tanh_backward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:91:63
[INFO] [stdout]     |
[INFO] [stdout] 91  |                 match result.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:414:21
[INFO] [stdout]     |
[INFO] [stdout] 414 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:106:20
[INFO] [stdout]     |
[INFO] [stdout] 106 |                 Ok(try!(match CUDNN.sigmoid_backward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:416:21
[INFO] [stdout]     |
[INFO] [stdout] 416 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x_diff, self.device()) }), //src_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:108:21
[INFO] [stdout]     |
[INFO] [stdout] 108 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:418:21
[INFO] [stdout]     |
[INFO] [stdout] 418 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(result, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:110:21
[INFO] [stdout]     |
[INFO] [stdout] 110 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x_diff, self.device()) }), //src_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:420:21
[INFO] [stdout]     |
[INFO] [stdout] 420 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result_diff, self.device()) }), // dest_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:112:21
[INFO] [stdout]     |
[INFO] [stdout] 112 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(result, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:413:22
[INFO] [stdout]     |
[INFO] [stdout] 413 |                     &try!(x.cudnn_tensor_desc_flat()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:114:21
[INFO] [stdout]     |
[INFO] [stdout] 114 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result_diff, self.device()) }), // dest_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:415:22
[INFO] [stdout]     |
[INFO] [stdout] 415 |                     &try!(x_diff.cudnn_tensor_desc_flat()), // src_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:107:22
[INFO] [stdout]     |
[INFO] [stdout] 107 |                     &try!(x.cudnn_tensor_desc_flat()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:417:22
[INFO] [stdout]     |
[INFO] [stdout] 417 |                     &try!(result.cudnn_tensor_desc_flat()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:109:22
[INFO] [stdout]     |
[INFO] [stdout] 109 |                     &try!(x_diff.cudnn_tensor_desc_flat()), // src_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:419:22
[INFO] [stdout]     |
[INFO] [stdout] 419 |                     &try!(result_diff.cudnn_tensor_desc_flat()), // dest_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:111:22
[INFO] [stdout]     |
[INFO] [stdout] 111 |                     &try!(result.cudnn_tensor_desc_flat()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:651:58
[INFO] [stdout]     |
[INFO] [stdout] 651 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:113:22
[INFO] [stdout]     |
[INFO] [stdout] 113 |                     &try!(result_diff.cudnn_tensor_desc_flat()), // dest_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:311:1
[INFO] [stdout]     |
[INFO] [stdout] 311 | impl_ops_sigmoid_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:664:20
[INFO] [stdout]     |
[INFO] [stdout] 664 |                 Ok(try!(match CUDNN.softmax_forward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:208:58
[INFO] [stdout]     |
[INFO] [stdout] 208 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:666:21
[INFO] [stdout]     |
[INFO] [stdout] 666 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:221:20
[INFO] [stdout]     |
[INFO] [stdout] 221 |                 Ok(try!(match CUDNN.relu_forward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:668:21
[INFO] [stdout]     |
[INFO] [stdout] 668 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:223:21
[INFO] [stdout]     |
[INFO] [stdout] 223 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:665:22
[INFO] [stdout]     |
[INFO] [stdout] 665 |                     &try!(x.cudnn_tensor_desc_softmax()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:225:21
[INFO] [stdout]     |
[INFO] [stdout] 225 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:667:22
[INFO] [stdout]     |
[INFO] [stdout] 667 |                     &try!(result.cudnn_tensor_desc_softmax()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:222:22
[INFO] [stdout]     |
[INFO] [stdout] 222 |                     &try!(x.cudnn_tensor_desc_flat()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:684:58
[INFO] [stdout]     |
[INFO] [stdout] 684 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:224:22
[INFO] [stdout]     |
[INFO] [stdout] 224 |                     &try!(result.cudnn_tensor_desc_flat()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:685:63
[INFO] [stdout]     |
[INFO] [stdout] 685 |                 match x_diff.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:242:58
[INFO] [stdout]     |
[INFO] [stdout] 242 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:699:20
[INFO] [stdout]     |
[INFO] [stdout] 699 |                 Ok(try!(match CUDNN.softmax_backward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:243:63
[INFO] [stdout]     |
[INFO] [stdout] 243 |                 match x_diff.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:701:21
[INFO] [stdout]     |
[INFO] [stdout] 701 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:244:63
[INFO] [stdout]     |
[INFO] [stdout] 244 |                 match result.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:703:21
[INFO] [stdout]     |
[INFO] [stdout] 703 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x_diff, self.device()) }), //src_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:259:20
[INFO] [stdout]     |
[INFO] [stdout] 259 |                 Ok(try!(match CUDNN.relu_backward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:705:21
[INFO] [stdout]     |
[INFO] [stdout] 705 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result_diff, self.device()) }), // dest_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:261:21
[INFO] [stdout]     |
[INFO] [stdout] 261 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:700:22
[INFO] [stdout]     |
[INFO] [stdout] 700 |                     &try!(x.cudnn_tensor_desc_softmax()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:263:21
[INFO] [stdout]     |
[INFO] [stdout] 263 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x_diff, self.device()) }), //src_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:702:22
[INFO] [stdout]     |
[INFO] [stdout] 702 |                     &try!(x_diff.cudnn_tensor_desc_softmax()), // src_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:265:21
[INFO] [stdout]     |
[INFO] [stdout] 265 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(result, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:704:22
[INFO] [stdout]     |
[INFO] [stdout] 704 |                     &try!(result_diff.cudnn_tensor_desc_softmax()), // dest_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:267:21
[INFO] [stdout]     |
[INFO] [stdout] 267 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result_diff, self.device()) }), // dest_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:727:58
[INFO] [stdout]     |
[INFO] [stdout] 727 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:260:22
[INFO] [stdout]     |
[INFO] [stdout] 260 |                     &try!(x.cudnn_tensor_desc_flat()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:740:20
[INFO] [stdout]     |
[INFO] [stdout] 740 |                 Ok(try!(match CUDNN.log_softmax_forward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:262:22
[INFO] [stdout]     |
[INFO] [stdout] 262 |                     &try!(x_diff.cudnn_tensor_desc_flat()), // src_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:742:21
[INFO] [stdout]     |
[INFO] [stdout] 742 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:264:22
[INFO] [stdout]     |
[INFO] [stdout] 264 |                     &try!(result.cudnn_tensor_desc_flat()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:744:21
[INFO] [stdout]     |
[INFO] [stdout] 744 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:266:22
[INFO] [stdout]     |
[INFO] [stdout] 266 |                     &try!(result_diff.cudnn_tensor_desc_flat()), // dest_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:312:1
[INFO] [stdout]     |
[INFO] [stdout] 312 | impl_ops_relu_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:741:22
[INFO] [stdout]     |
[INFO] [stdout] 741 |                     &try!(x.cudnn_tensor_desc_softmax()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:361:58
[INFO] [stdout]     |
[INFO] [stdout] 361 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:743:22
[INFO] [stdout]     |
[INFO] [stdout] 743 |                     &try!(result.cudnn_tensor_desc_softmax()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:374:20
[INFO] [stdout]     |
[INFO] [stdout] 374 |                 Ok(try!(match CUDNN.tanh_forward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:760:58
[INFO] [stdout]     |
[INFO] [stdout] 760 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:376:21
[INFO] [stdout]     |
[INFO] [stdout] 376 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:761:63
[INFO] [stdout]     |
[INFO] [stdout] 761 |                 match x_diff.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:775:20
[INFO] [stdout]     |
[INFO] [stdout] 775 |                 Ok(try!(match CUDNN.log_softmax_backward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:378:21
[INFO] [stdout]     |
[INFO] [stdout] 378 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:777:21
[INFO] [stdout]     |
[INFO] [stdout] 777 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:375:22
[INFO] [stdout]     |
[INFO] [stdout] 375 |                     &try!(x.cudnn_tensor_desc_flat()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:779:21
[INFO] [stdout]     |
[INFO] [stdout] 779 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x_diff, self.device()) }), //src_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:377:22
[INFO] [stdout]     |
[INFO] [stdout] 377 |                     &try!(result.cudnn_tensor_desc_flat()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:781:21
[INFO] [stdout]     |
[INFO] [stdout] 781 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result_diff, self.device()) }), // dest_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:776:22
[INFO] [stdout]     |
[INFO] [stdout] 776 |                     &try!(x.cudnn_tensor_desc_softmax()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:778:22
[INFO] [stdout]     |
[INFO] [stdout] 778 |                     &try!(x_diff.cudnn_tensor_desc_softmax()), // src_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:780:22
[INFO] [stdout]     |
[INFO] [stdout] 780 |                     &try!(result_diff.cudnn_tensor_desc_softmax()), // dest_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:814:58
[INFO] [stdout]     |
[INFO] [stdout] 814 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:828:20
[INFO] [stdout]     |
[INFO] [stdout] 828 |                 Ok(try!(match CUDNN.lrn_forward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:831:21
[INFO] [stdout]     |
[INFO] [stdout] 831 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:833:21
[INFO] [stdout]     |
[INFO] [stdout] 833 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:830:22
[INFO] [stdout]     |
[INFO] [stdout] 830 |                     &try!(x.cudnn_tensor_desc()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:395:58
[INFO] [stdout]     |
[INFO] [stdout] 395 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:832:22
[INFO] [stdout]     |
[INFO] [stdout] 832 |                     &try!(result.cudnn_tensor_desc()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:396:63
[INFO] [stdout]     |
[INFO] [stdout] 396 |                 match x_diff.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:852:58
[INFO] [stdout]     |
[INFO] [stdout] 852 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:397:63
[INFO] [stdout]     |
[INFO] [stdout] 397 |                 match result.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:853:63
[INFO] [stdout]     |
[INFO] [stdout] 853 |                 match x_diff.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:412:20
[INFO] [stdout]     |
[INFO] [stdout] 412 |                 Ok(try!(match CUDNN.tanh_backward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:854:63
[INFO] [stdout]     |
[INFO] [stdout] 854 |                 match result.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:414:21
[INFO] [stdout]     |
[INFO] [stdout] 414 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:871:20
[INFO] [stdout]     |
[INFO] [stdout] 871 |                 Ok(try!(match CUDNN.lrn_backward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:416:21
[INFO] [stdout]     |
[INFO] [stdout] 416 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x_diff, self.device()) }), //src_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:874:21
[INFO] [stdout]     |
[INFO] [stdout] 874 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:418:21
[INFO] [stdout]     |
[INFO] [stdout] 418 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(result, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:876:21
[INFO] [stdout]     |
[INFO] [stdout] 876 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x_diff, self.device()) }), //src_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:420:21
[INFO] [stdout]     |
[INFO] [stdout] 420 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result_diff, self.device()) }), // dest_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:878:21
[INFO] [stdout]     |
[INFO] [stdout] 878 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(result, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:880:21
[INFO] [stdout]     |
[INFO] [stdout] 880 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result_diff, self.device()) }), // dest_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:413:22
[INFO] [stdout]     |
[INFO] [stdout] 413 |                     &try!(x.cudnn_tensor_desc_flat()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:415:22
[INFO] [stdout]     |
[INFO] [stdout] 415 |                     &try!(x_diff.cudnn_tensor_desc_flat()), // src_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:873:22
[INFO] [stdout]     |
[INFO] [stdout] 873 |                     &try!(x.cudnn_tensor_desc()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:417:22
[INFO] [stdout]     |
[INFO] [stdout] 417 |                     &try!(result.cudnn_tensor_desc_flat()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:875:22
[INFO] [stdout]     |
[INFO] [stdout] 875 |                     &try!(x_diff.cudnn_tensor_desc()), // src_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:419:22
[INFO] [stdout]     |
[INFO] [stdout] 419 |                     &try!(result_diff.cudnn_tensor_desc_flat()), // dest_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:313:1
[INFO] [stdout]     |
[INFO] [stdout] 313 | impl_ops_tanh_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | -------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:877:22
[INFO] [stdout]     |
[INFO] [stdout] 877 |                     &try!(result.cudnn_tensor_desc()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:651:58
[INFO] [stdout]     |
[INFO] [stdout] 651 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:879:22
[INFO] [stdout]     |
[INFO] [stdout] 879 |                     &try!(result_diff.cudnn_tensor_desc()), // dest_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:664:20
[INFO] [stdout]     |
[INFO] [stdout] 664 |                 Ok(try!(match CUDNN.softmax_forward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:914:58
[INFO] [stdout]     |
[INFO] [stdout] 914 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:666:21
[INFO] [stdout]     |
[INFO] [stdout] 666 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:928:20
[INFO] [stdout]     |
[INFO] [stdout] 928 |                 Ok(try!(match CUDNN.pooling_max_forward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:931:21
[INFO] [stdout]     |
[INFO] [stdout] 931 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:668:21
[INFO] [stdout]     |
[INFO] [stdout] 668 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:933:21
[INFO] [stdout]     |
[INFO] [stdout] 933 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:665:22
[INFO] [stdout]     |
[INFO] [stdout] 665 |                     &try!(x.cudnn_tensor_desc_softmax()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:930:22
[INFO] [stdout]     |
[INFO] [stdout] 930 |                     &try!(x.cudnn_tensor_desc()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:667:22
[INFO] [stdout]     |
[INFO] [stdout] 667 |                     &try!(result.cudnn_tensor_desc_softmax()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:932:22
[INFO] [stdout]     |
[INFO] [stdout] 932 |                     &try!(result.cudnn_tensor_desc()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:684:58
[INFO] [stdout]     |
[INFO] [stdout] 684 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:952:58
[INFO] [stdout]     |
[INFO] [stdout] 952 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:685:63
[INFO] [stdout]     |
[INFO] [stdout] 685 |                 match x_diff.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:953:63
[INFO] [stdout]     |
[INFO] [stdout] 953 |                 match x_diff.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:699:20
[INFO] [stdout]     |
[INFO] [stdout] 699 |                 Ok(try!(match CUDNN.softmax_backward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:954:63
[INFO] [stdout]     |
[INFO] [stdout] 954 |                 match result.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:701:21
[INFO] [stdout]     |
[INFO] [stdout] 701 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:971:20
[INFO] [stdout]     |
[INFO] [stdout] 971 |                 Ok(try!(match CUDNN.pooling_max_backward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:703:21
[INFO] [stdout]     |
[INFO] [stdout] 703 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x_diff, self.device()) }), //src_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:974:21
[INFO] [stdout]     |
[INFO] [stdout] 974 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:705:21
[INFO] [stdout]     |
[INFO] [stdout] 705 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result_diff, self.device()) }), // dest_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:976:21
[INFO] [stdout]     |
[INFO] [stdout] 976 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x_diff, self.device()) }), //src_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:700:22
[INFO] [stdout]     |
[INFO] [stdout] 700 |                     &try!(x.cudnn_tensor_desc_softmax()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:702:22
[INFO] [stdout]     |
[INFO] [stdout] 702 |                     &try!(x_diff.cudnn_tensor_desc_softmax()), // src_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:978:21
[INFO] [stdout]     |
[INFO] [stdout] 978 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(result, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:704:22
[INFO] [stdout]     |
[INFO] [stdout] 704 |                     &try!(result_diff.cudnn_tensor_desc_softmax()), // dest_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:314:1
[INFO] [stdout]     |
[INFO] [stdout] 314 | impl_ops_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:980:21
[INFO] [stdout]     |
[INFO] [stdout] 980 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result_diff, self.device()) }), // dest_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:727:58
[INFO] [stdout]     |
[INFO] [stdout] 727 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:973:22
[INFO] [stdout]     |
[INFO] [stdout] 973 |                     &try!(x.cudnn_tensor_desc()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:740:20
[INFO] [stdout]     |
[INFO] [stdout] 740 |                 Ok(try!(match CUDNN.log_softmax_forward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:975:22
[INFO] [stdout]     |
[INFO] [stdout] 975 |                     &try!(x_diff.cudnn_tensor_desc()), // src_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:742:21
[INFO] [stdout]     |
[INFO] [stdout] 742 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:977:22
[INFO] [stdout]     |
[INFO] [stdout] 977 |                     &try!(result.cudnn_tensor_desc()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:744:21
[INFO] [stdout]     |
[INFO] [stdout] 744 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:979:22
[INFO] [stdout]     |
[INFO] [stdout] 979 |                     &try!(result_diff.cudnn_tensor_desc()), // dest_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:741:22
[INFO] [stdout]     |
[INFO] [stdout] 741 |                     &try!(x.cudnn_tensor_desc_softmax()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:135:58
[INFO] [stdout]     |
[INFO] [stdout] 135 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:146:20
[INFO] [stdout]     |
[INFO] [stdout] 146 |                 Ok(try!(match CUDNN.sigmoid_forward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:743:22
[INFO] [stdout]     |
[INFO] [stdout] 743 |                     &try!(result.cudnn_tensor_desc_softmax()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:148:21
[INFO] [stdout]     |
[INFO] [stdout] 148 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:760:58
[INFO] [stdout]     |
[INFO] [stdout] 760 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:150:21
[INFO] [stdout]     |
[INFO] [stdout] 150 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(x, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:761:63
[INFO] [stdout]     |
[INFO] [stdout] 761 |                 match x_diff.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:147:22
[INFO] [stdout]     |
[INFO] [stdout] 147 |                     &try!(x.cudnn_tensor_desc_flat()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:775:20
[INFO] [stdout]     |
[INFO] [stdout] 775 |                 Ok(try!(match CUDNN.log_softmax_backward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:149:22
[INFO] [stdout]     |
[INFO] [stdout] 149 |                     &try!(x.cudnn_tensor_desc_flat()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:777:21
[INFO] [stdout]     |
[INFO] [stdout] 777 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:779:21
[INFO] [stdout]     |
[INFO] [stdout] 779 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x_diff, self.device()) }), //src_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:165:58
[INFO] [stdout]     |
[INFO] [stdout] 165 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:781:21
[INFO] [stdout]     |
[INFO] [stdout] 781 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result_diff, self.device()) }), // dest_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:166:63
[INFO] [stdout]     |
[INFO] [stdout] 166 |                 match x_diff.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:776:22
[INFO] [stdout]     |
[INFO] [stdout] 776 |                     &try!(x.cudnn_tensor_desc_softmax()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:178:20
[INFO] [stdout]     |
[INFO] [stdout] 178 |                 Ok(try!(match CUDNN.sigmoid_backward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:778:22
[INFO] [stdout]     |
[INFO] [stdout] 778 |                     &try!(x_diff.cudnn_tensor_desc_softmax()), // src_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:180:21
[INFO] [stdout]     |
[INFO] [stdout] 180 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:780:22
[INFO] [stdout]     |
[INFO] [stdout] 780 |                     &try!(result_diff.cudnn_tensor_desc_softmax()), // dest_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:315:1
[INFO] [stdout]     |
[INFO] [stdout] 315 | impl_ops_log_softmax_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_log_softmax_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:182:21
[INFO] [stdout]     |
[INFO] [stdout] 182 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x_diff, self.device()) }), //src_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:184:21
[INFO] [stdout]     |
[INFO] [stdout] 184 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:814:58
[INFO] [stdout]     |
[INFO] [stdout] 814 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:828:20
[INFO] [stdout]     |
[INFO] [stdout] 828 |                 Ok(try!(match CUDNN.lrn_forward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:186:21
[INFO] [stdout]     |
[INFO] [stdout] 186 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(x_diff, self.device()) }), // dest_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:831:21
[INFO] [stdout]     |
[INFO] [stdout] 831 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:179:22
[INFO] [stdout]     |
[INFO] [stdout] 179 |                     &try!(x.cudnn_tensor_desc_flat()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:181:22
[INFO] [stdout]     |
[INFO] [stdout] 181 |                     &try!(x_diff.cudnn_tensor_desc_flat()), // src_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:833:21
[INFO] [stdout]     |
[INFO] [stdout] 833 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:183:22
[INFO] [stdout]     |
[INFO] [stdout] 183 |                     &try!(x.cudnn_tensor_desc_flat()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:830:22
[INFO] [stdout]     |
[INFO] [stdout] 830 |                     &try!(x.cudnn_tensor_desc()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:185:22
[INFO] [stdout]     |
[INFO] [stdout] 185 |                     &try!(x_diff.cudnn_tensor_desc_flat()), // dest_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:832:22
[INFO] [stdout]     |
[INFO] [stdout] 832 |                     &try!(result.cudnn_tensor_desc()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:288:58
[INFO] [stdout]     |
[INFO] [stdout] 288 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:299:20
[INFO] [stdout]     |
[INFO] [stdout] 299 |                 Ok(try!(match CUDNN.relu_forward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:852:58
[INFO] [stdout]     |
[INFO] [stdout] 852 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:301:21
[INFO] [stdout]     |
[INFO] [stdout] 301 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:853:63
[INFO] [stdout]     |
[INFO] [stdout] 853 |                 match x_diff.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:854:63
[INFO] [stdout]     |
[INFO] [stdout] 854 |                 match result.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:303:21
[INFO] [stdout]     |
[INFO] [stdout] 303 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(x, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:871:20
[INFO] [stdout]     |
[INFO] [stdout] 871 |                 Ok(try!(match CUDNN.lrn_backward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:300:22
[INFO] [stdout]     |
[INFO] [stdout] 300 |                     &try!(x.cudnn_tensor_desc_flat()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:874:21
[INFO] [stdout]     |
[INFO] [stdout] 874 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:302:22
[INFO] [stdout]     |
[INFO] [stdout] 302 |                     &try!(x.cudnn_tensor_desc_flat()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:876:21
[INFO] [stdout]     |
[INFO] [stdout] 876 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x_diff, self.device()) }), //src_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:878:21
[INFO] [stdout]     |
[INFO] [stdout] 878 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(result, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:318:58
[INFO] [stdout]     |
[INFO] [stdout] 318 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:880:21
[INFO] [stdout]     |
[INFO] [stdout] 880 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result_diff, self.device()) }), // dest_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:319:63
[INFO] [stdout]     |
[INFO] [stdout] 319 |                 match x_diff.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:873:22
[INFO] [stdout]     |
[INFO] [stdout] 873 |                     &try!(x.cudnn_tensor_desc()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:331:20
[INFO] [stdout]     |
[INFO] [stdout] 331 |                 Ok(try!(match CUDNN.relu_backward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:875:22
[INFO] [stdout]     |
[INFO] [stdout] 875 |                     &try!(x_diff.cudnn_tensor_desc()), // src_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:333:21
[INFO] [stdout]     |
[INFO] [stdout] 333 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:877:22
[INFO] [stdout]     |
[INFO] [stdout] 877 |                     &try!(result.cudnn_tensor_desc()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:335:21
[INFO] [stdout]     |
[INFO] [stdout] 335 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x_diff, self.device()) }), //src_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:879:22
[INFO] [stdout]     |
[INFO] [stdout] 879 |                     &try!(result_diff.cudnn_tensor_desc()), // dest_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:316:1
[INFO] [stdout]     |
[INFO] [stdout] 316 | impl_ops_lrn_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_lrn_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:337:21
[INFO] [stdout]     |
[INFO] [stdout] 337 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:914:58
[INFO] [stdout]     |
[INFO] [stdout] 914 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:339:21
[INFO] [stdout]     |
[INFO] [stdout] 339 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(x_diff, self.device()) }), // dest_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:928:20
[INFO] [stdout]     |
[INFO] [stdout] 928 |                 Ok(try!(match CUDNN.pooling_max_forward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:332:22
[INFO] [stdout]     |
[INFO] [stdout] 332 |                     &try!(x.cudnn_tensor_desc_flat()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:931:21
[INFO] [stdout]     |
[INFO] [stdout] 931 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:334:22
[INFO] [stdout]     |
[INFO] [stdout] 334 |                     &try!(x_diff.cudnn_tensor_desc_flat()), // src_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:933:21
[INFO] [stdout]     |
[INFO] [stdout] 933 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:336:22
[INFO] [stdout]     |
[INFO] [stdout] 336 |                     &try!(x.cudnn_tensor_desc_flat()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:338:22
[INFO] [stdout]     |
[INFO] [stdout] 338 |                     &try!(x_diff.cudnn_tensor_desc_flat()), // dest_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:930:22
[INFO] [stdout]     |
[INFO] [stdout] 930 |                     &try!(x.cudnn_tensor_desc()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:441:58
[INFO] [stdout]     |
[INFO] [stdout] 441 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:321:1
[INFO] [stdout]     |
[INFO] [stdout] 321 | impl_ops_tanh_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:932:22
[INFO] [stdout]     |
[INFO] [stdout] 932 |                     &try!(result.cudnn_tensor_desc()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:452:20
[INFO] [stdout]     |
[INFO] [stdout] 452 |                 Ok(try!(match CUDNN.tanh_forward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:321:1
[INFO] [stdout]     |
[INFO] [stdout] 321 | impl_ops_tanh_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:952:58
[INFO] [stdout]     |
[INFO] [stdout] 952 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:454:21
[INFO] [stdout]     |
[INFO] [stdout] 454 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:321:1
[INFO] [stdout]     |
[INFO] [stdout] 321 | impl_ops_tanh_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:953:63
[INFO] [stdout]     |
[INFO] [stdout] 953 |                 match x_diff.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:456:21
[INFO] [stdout]     |
[INFO] [stdout] 456 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(x, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:321:1
[INFO] [stdout]     |
[INFO] [stdout] 321 | impl_ops_tanh_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:954:63
[INFO] [stdout]     |
[INFO] [stdout] 954 |                 match result.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:453:22
[INFO] [stdout]     |
[INFO] [stdout] 453 |                     &try!(x.cudnn_tensor_desc_flat()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:321:1
[INFO] [stdout]     |
[INFO] [stdout] 321 | impl_ops_tanh_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:971:20
[INFO] [stdout]     |
[INFO] [stdout] 971 |                 Ok(try!(match CUDNN.pooling_max_backward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:455:22
[INFO] [stdout]     |
[INFO] [stdout] 455 |                     &try!(x.cudnn_tensor_desc_flat()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:321:1
[INFO] [stdout]     |
[INFO] [stdout] 321 | impl_ops_tanh_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:974:21
[INFO] [stdout]     |
[INFO] [stdout] 974 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:976:21
[INFO] [stdout]     |
[INFO] [stdout] 976 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x_diff, self.device()) }), //src_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:471:58
[INFO] [stdout]     |
[INFO] [stdout] 471 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:321:1
[INFO] [stdout]     |
[INFO] [stdout] 321 | impl_ops_tanh_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:978:21
[INFO] [stdout]     |
[INFO] [stdout] 978 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(result, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:472:63
[INFO] [stdout]     |
[INFO] [stdout] 472 |                 match x_diff.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:321:1
[INFO] [stdout]     |
[INFO] [stdout] 321 | impl_ops_tanh_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:980:21
[INFO] [stdout]     |
[INFO] [stdout] 980 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result_diff, self.device()) }), // dest_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:484:20
[INFO] [stdout]     |
[INFO] [stdout] 484 |                 Ok(try!(match CUDNN.tanh_backward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:321:1
[INFO] [stdout]     |
[INFO] [stdout] 321 | impl_ops_tanh_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:973:22
[INFO] [stdout]     |
[INFO] [stdout] 973 |                     &try!(x.cudnn_tensor_desc()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:486:21
[INFO] [stdout]     |
[INFO] [stdout] 486 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:321:1
[INFO] [stdout]     |
[INFO] [stdout] 321 | impl_ops_tanh_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:975:22
[INFO] [stdout]     |
[INFO] [stdout] 975 |                     &try!(x_diff.cudnn_tensor_desc()), // src_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:488:21
[INFO] [stdout]     |
[INFO] [stdout] 488 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x_diff, self.device()) }), //src_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:321:1
[INFO] [stdout]     |
[INFO] [stdout] 321 | impl_ops_tanh_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:977:22
[INFO] [stdout]     |
[INFO] [stdout] 977 |                     &try!(result.cudnn_tensor_desc()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:979:22
[INFO] [stdout]     |
[INFO] [stdout] 979 |                     &try!(result_diff.cudnn_tensor_desc()), // dest_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:317:1
[INFO] [stdout]     |
[INFO] [stdout] 317 | impl_ops_pooling_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ----------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_pooling_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:490:21
[INFO] [stdout]     |
[INFO] [stdout] 490 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:321:1
[INFO] [stdout]     |
[INFO] [stdout] 321 | impl_ops_tanh_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:135:58
[INFO] [stdout]     |
[INFO] [stdout] 135 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:492:21
[INFO] [stdout]     |
[INFO] [stdout] 492 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(x_diff, self.device()) }), // dest_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:321:1
[INFO] [stdout]     |
[INFO] [stdout] 321 | impl_ops_tanh_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:146:20
[INFO] [stdout]     |
[INFO] [stdout] 146 |                 Ok(try!(match CUDNN.sigmoid_forward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:485:22
[INFO] [stdout]     |
[INFO] [stdout] 485 |                     &try!(x.cudnn_tensor_desc_flat()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:321:1
[INFO] [stdout]     |
[INFO] [stdout] 321 | impl_ops_tanh_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:487:22
[INFO] [stdout]     |
[INFO] [stdout] 487 |                     &try!(x_diff.cudnn_tensor_desc_flat()), // src_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:321:1
[INFO] [stdout]     |
[INFO] [stdout] 321 | impl_ops_tanh_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:148:21
[INFO] [stdout]     |
[INFO] [stdout] 148 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:489:22
[INFO] [stdout]     |
[INFO] [stdout] 489 |                     &try!(x.cudnn_tensor_desc_flat()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:321:1
[INFO] [stdout]     |
[INFO] [stdout] 321 | impl_ops_tanh_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:150:21
[INFO] [stdout]     |
[INFO] [stdout] 150 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(x, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:491:22
[INFO] [stdout]     |
[INFO] [stdout] 491 |                     &try!(x_diff.cudnn_tensor_desc_flat()), // dest_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:321:1
[INFO] [stdout]     |
[INFO] [stdout] 321 | impl_ops_tanh_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_tanh_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:147:22
[INFO] [stdout]     |
[INFO] [stdout] 147 |                     &try!(x.cudnn_tensor_desc_flat()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:263:32
[INFO] [stdout]     |
[INFO] [stdout] 263 |                 let src_desc = try!(src.cudnn_tensor_desc());
[INFO] [stdout]     |                                ^^^
[INFO] [stdout] ...
[INFO] [stdout] 332 | impl_convolution_for_cuda_backend!(f64, ::cudnn::utils::DataType::Double);
[INFO] [stdout]     | ------------------------------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:149:22
[INFO] [stdout]     |
[INFO] [stdout] 149 |                     &try!(x.cudnn_tensor_desc_flat()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:264:33
[INFO] [stdout]     |
[INFO] [stdout] 264 |                 let dest_desc = try!(dest.cudnn_tensor_desc());
[INFO] [stdout]     |                                 ^^^
[INFO] [stdout] ...
[INFO] [stdout] 332 | impl_convolution_for_cuda_backend!(f64, ::cudnn::utils::DataType::Double);
[INFO] [stdout]     | ------------------------------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:165:58
[INFO] [stdout]     |
[INFO] [stdout] 165 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:265:35
[INFO] [stdout]     |
[INFO] [stdout] 265 |                 let filter_desc = try!(filter.cudnn_filter_desc());
[INFO] [stdout]     |                                   ^^^
[INFO] [stdout] ...
[INFO] [stdout] 332 | impl_convolution_for_cuda_backend!(f64, ::cudnn::utils::DataType::Double);
[INFO] [stdout]     | ------------------------------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:166:63
[INFO] [stdout]     |
[INFO] [stdout] 166 |                 match x_diff.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error: unnecessary qualification
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:332:41
[INFO] [stdout]     |
[INFO] [stdout] 332 | impl_convolution_for_cuda_backend!(f64, ::cudnn::utils::DataType::Double);
[INFO] [stdout]     |                                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]     |
[INFO] [stdout] help: remove the unnecessary path segments
[INFO] [stdout]     |
[INFO] [stdout] 332 - impl_convolution_for_cuda_backend!(f64, ::cudnn::utils::DataType::Double);
[INFO] [stdout] 332 + impl_convolution_for_cuda_backend!(f64, utils::DataType::Double);
[INFO] [stdout]     |
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:268:40
[INFO] [stdout]     |
[INFO] [stdout] 268 |                 let useable_algo_fwd = try!(algo_fwd.find_cudnn_algo(&filter_desc, &conv_desc, &src_desc, &dest_desc));
[INFO] [stdout]     |                                        ^^^
[INFO] [stdout] ...
[INFO] [stdout] 332 | impl_convolution_for_cuda_backend!(f64, ::cudnn::utils::DataType::Double);
[INFO] [stdout]     | ------------------------------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:178:20
[INFO] [stdout]     |
[INFO] [stdout] 178 |                 Ok(try!(match CUDNN.sigmoid_backward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:180:21
[INFO] [stdout]     |
[INFO] [stdout] 180 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:269:47
[INFO] [stdout]     |
[INFO] [stdout] 269 |                 let useable_algo_bwd_filter = try!(algo_bwd_filter.find_cudnn_algo(&filter_desc, &conv_desc, &src_desc, &dest_desc));
[INFO] [stdout]     |                                               ^^^
[INFO] [stdout] ...
[INFO] [stdout] 332 | impl_convolution_for_cuda_backend!(f64, ::cudnn::utils::DataType::Double);
[INFO] [stdout]     | ------------------------------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:182:21
[INFO] [stdout]     |
[INFO] [stdout] 182 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x_diff, self.device()) }), //src_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/mod.rs:270:45
[INFO] [stdout]     |
[INFO] [stdout] 270 |                 let useable_algo_bwd_data = try!(algo_bwd_data.find_cudnn_algo(&filter_desc, &conv_desc, &src_desc, &dest_desc));
[INFO] [stdout]     |                                             ^^^
[INFO] [stdout] ...
[INFO] [stdout] 332 | impl_convolution_for_cuda_backend!(f64, ::cudnn::utils::DataType::Double);
[INFO] [stdout]     | ------------------------------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:184:21
[INFO] [stdout]     |
[INFO] [stdout] 184 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:186:21
[INFO] [stdout]     |
[INFO] [stdout] 186 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(x_diff, self.device()) }), // dest_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:516:54
[INFO] [stdout]     |
[INFO] [stdout] 516 |             match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:332:1
[INFO] [stdout]     |
[INFO] [stdout] 332 | impl_convolution_for_cuda_backend!(f64, ::cudnn::utils::DataType::Double);
[INFO] [stdout]     | ------------------------------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:179:22
[INFO] [stdout]     |
[INFO] [stdout] 179 |                     &try!(x.cudnn_tensor_desc_flat()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:533:16
[INFO] [stdout]     |
[INFO] [stdout] 533 |             Ok(try!(match CUDNN.convolution_forward(
[INFO] [stdout]     |                ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:332:1
[INFO] [stdout]     |
[INFO] [stdout] 332 | impl_convolution_for_cuda_backend!(f64, ::cudnn::utils::DataType::Double);
[INFO] [stdout]     | ------------------------------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:181:22
[INFO] [stdout]     |
[INFO] [stdout] 181 |                     &try!(x_diff.cudnn_tensor_desc_flat()), // src_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:183:22
[INFO] [stdout]     |
[INFO] [stdout] 183 |                     &try!(x.cudnn_tensor_desc_flat()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:535:17
[INFO] [stdout]     |
[INFO] [stdout] 535 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(workspace, self.device()) }),
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:332:1
[INFO] [stdout]     |
[INFO] [stdout] 332 | impl_convolution_for_cuda_backend!(f64, ::cudnn::utils::DataType::Double);
[INFO] [stdout]     | ------------------------------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:185:22
[INFO] [stdout]     |
[INFO] [stdout] 185 |                     &try!(x_diff.cudnn_tensor_desc_flat()), // dest_diff_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:319:1
[INFO] [stdout]     |
[INFO] [stdout] 319 | impl_ops_sigmoid_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | --------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_sigmoid_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:536:17
[INFO] [stdout]     |
[INFO] [stdout] 536 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(filter, self.device()) }),
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:332:1
[INFO] [stdout]     |
[INFO] [stdout] 332 | impl_convolution_for_cuda_backend!(f64, ::cudnn::utils::DataType::Double);
[INFO] [stdout]     | ------------------------------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:288:58
[INFO] [stdout]     |
[INFO] [stdout] 288 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:299:20
[INFO] [stdout]     |
[INFO] [stdout] 299 |                 Ok(try!(match CUDNN.relu_forward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:538:17
[INFO] [stdout]     |
[INFO] [stdout] 538 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:332:1
[INFO] [stdout]     |
[INFO] [stdout] 332 | impl_convolution_for_cuda_backend!(f64, ::cudnn::utils::DataType::Double);
[INFO] [stdout]     | ------------------------------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:540:17
[INFO] [stdout]     |
[INFO] [stdout] 540 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(result, self.device()) }), // dest_data
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:332:1
[INFO] [stdout]     |
[INFO] [stdout] 332 | impl_convolution_for_cuda_backend!(f64, ::cudnn::utils::DataType::Double);
[INFO] [stdout]     | ------------------------------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:301:21
[INFO] [stdout]     |
[INFO] [stdout] 301 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:303:21
[INFO] [stdout]     |
[INFO] [stdout] 303 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(x, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:537:18
[INFO] [stdout]     |
[INFO] [stdout] 537 |                 &try!(x.cudnn_tensor_desc()), // src_desc
[INFO] [stdout]     |                  ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:332:1
[INFO] [stdout]     |
[INFO] [stdout] 332 | impl_convolution_for_cuda_backend!(f64, ::cudnn::utils::DataType::Double);
[INFO] [stdout]     | ------------------------------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:300:22
[INFO] [stdout]     |
[INFO] [stdout] 300 |                     &try!(x.cudnn_tensor_desc_flat()), // src_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:539:18
[INFO] [stdout]     |
[INFO] [stdout] 539 |                 &try!(result.cudnn_tensor_desc()), // dest_desc
[INFO] [stdout]     |                  ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:332:1
[INFO] [stdout]     |
[INFO] [stdout] 332 | impl_convolution_for_cuda_backend!(f64, ::cudnn::utils::DataType::Double);
[INFO] [stdout]     | ------------------------------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:302:22
[INFO] [stdout]     |
[INFO] [stdout] 302 |                     &try!(x.cudnn_tensor_desc_flat()), // dest_desc
[INFO] [stdout]     |                      ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:559:61
[INFO] [stdout]     |
[INFO] [stdout] 559 |             match src_data.add_device(self.device()) { _ => try!(src_data.sync(self.device())) }
[INFO] [stdout]     |                                                             ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:332:1
[INFO] [stdout]     |
[INFO] [stdout] 332 | impl_convolution_for_cuda_backend!(f64, ::cudnn::utils::DataType::Double);
[INFO] [stdout]     | ------------------------------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:318:58
[INFO] [stdout]     |
[INFO] [stdout] 318 |                 match x.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                          ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:319:63
[INFO] [stdout]     |
[INFO] [stdout] 319 |                 match x_diff.add_device(self.device()) { _ => try!(x.sync(self.device())) }
[INFO] [stdout]     |                                                               ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:560:62
[INFO] [stdout]     |
[INFO] [stdout] 560 |             match dest_diff.add_device(self.device()) { _ => try!(dest_diff.sync(self.device())) }
[INFO] [stdout]     |                                                              ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:332:1
[INFO] [stdout]     |
[INFO] [stdout] 332 | impl_convolution_for_cuda_backend!(f64, ::cudnn::utils::DataType::Double);
[INFO] [stdout]     | ------------------------------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:331:20
[INFO] [stdout]     |
[INFO] [stdout] 331 |                 Ok(try!(match CUDNN.relu_backward(
[INFO] [stdout]     |                    ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:561:64
[INFO] [stdout]     |
[INFO] [stdout] 561 |             match filter_diff.add_device(self.device()) { _ => try!(filter_diff.sync(self.device())) }
[INFO] [stdout]     |                                                                ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:332:1
[INFO] [stdout]     |
[INFO] [stdout] 332 | impl_convolution_for_cuda_backend!(f64, ::cudnn::utils::DataType::Double);
[INFO] [stdout]     | ------------------------------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:333:21
[INFO] [stdout]     |
[INFO] [stdout] 333 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), //src_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:578:16
[INFO] [stdout]     |
[INFO] [stdout] 578 |             Ok(try!(match CUDNN.convolution_backward_filter(
[INFO] [stdout]     |                ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:332:1
[INFO] [stdout]     |
[INFO] [stdout] 332 | impl_convolution_for_cuda_backend!(f64, ::cudnn::utils::DataType::Double);
[INFO] [stdout]     | ------------------------------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:335:21
[INFO] [stdout]     |
[INFO] [stdout] 335 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x_diff, self.device()) }), //src_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:337:21
[INFO] [stdout]     |
[INFO] [stdout] 337 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr(x, self.device()) }), // dest_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[INFO] [stdout]     | ------------------------------------------------ in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_relu_pointwise_for` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:580:17
[INFO] [stdout]     |
[INFO] [stdout] 580 |                 try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(workspace, self.device()) }),
[INFO] [stdout]     |                 ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:332:1
[INFO] [stdout]     |
[INFO] [stdout] 332 | impl_convolution_for_cuda_backend!(f64, ::cudnn::utils::DataType::Double);
[INFO] [stdout]     | ------------------------------------------------------------------------- in this macro invocation
[INFO] [stdout]     |
[INFO] [stdout]     = note: this warning originates in the macro `impl_ops_convolution_for` which comes from the expansion of the macro `impl_convolution_for_cuda_backend` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: use of deprecated macro `try`: use the `?` operator instead
[INFO] [stdout]    --> src/frameworks/cuda/helper.rs:339:21
[INFO] [stdout]     |
[INFO] [stdout] 339 |                     try!(unsafe { ::frameworks::cuda::helper::receive_memory_ptr_mut(x_diff, self.device()) }), // dest_diff_data
[INFO] [stdout]     |                     ^^^
[INFO] [stdout]     |
[INFO] [stdout]    ::: src/frameworks/cuda/mod.rs:320:1
[INFO] [stdout]     |
[INFO] [stdout] 320 | impl_ops_relu_pointwise_for!(f32, Backend<Cuda>);
[WARN] too many lines in the log, truncating it
