[INFO] updating cached repository https://github.com/forgerpl/webcrawl
[INFO] running `"git" "-c" "credential.helper=" "-c" "credential.helper=/workspace/cargo-home/bin/git-credential-null" "-c" "remote.origin.fetch=refs/heads/*:refs/heads/*" "fetch" "origin" "--force" "--prune"`
[INFO] running `"git" "rev-parse" "HEAD"`
[INFO] [stdout] 29fd7499dd3bc8789b9dbc9b2f039fc704e67db7
[INFO] checking forgerpl/webcrawl against master#209b2be09fcaff937480d1fbbe8b31646e361c7a for pr-70917
[INFO] running `"git" "clone" "/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2Fforgerpl%2Fwebcrawl" "/workspace/builds/worker-6/source"`
[INFO] [stderr] Cloning into '/workspace/builds/worker-6/source'...
[INFO] [stderr] done.
[INFO] validating manifest of git repo https://github.com/forgerpl/webcrawl on toolchain 209b2be09fcaff937480d1fbbe8b31646e361c7a
[INFO] running `"/workspace/cargo-home/bin/cargo" "+209b2be09fcaff937480d1fbbe8b31646e361c7a" "read-manifest" "--manifest-path" "Cargo.toml"`
[INFO] started tweaking git repo https://github.com/forgerpl/webcrawl
[INFO] removed workspace from git repo https://github.com/forgerpl/webcrawl
[INFO] removed path dependency crawler from git repo https://github.com/forgerpl/webcrawl
[INFO] finished tweaking git repo https://github.com/forgerpl/webcrawl
[INFO] tweaked toml for git repo https://github.com/forgerpl/webcrawl written to /workspace/builds/worker-6/source/Cargo.toml
[INFO] crate git repo https://github.com/forgerpl/webcrawl already has a lockfile, it will not be regenerated
[INFO] running `"/workspace/cargo-home/bin/cargo" "+209b2be09fcaff937480d1fbbe8b31646e361c7a" "fetch" "--locked" "--manifest-path" "Cargo.toml"`
[INFO] [stderr] warning: dependency (crawler) specified without providing a local path, Git repository, or version to use. This will be considered an error in future versions
[INFO] [stderr]     Blocking waiting for file lock on package cache
[INFO] [stderr]     Updating crates.io index
[INFO] [stderr] error: the lock file /workspace/builds/worker-6/source/Cargo.lock needs to be updated but --locked was passed to prevent this
[INFO] [stderr] If you want to try to generate the lock file without accessing the network, use the --offline flag.
[INFO] the lockfile is outdated, regenerating it
[INFO] running `"/workspace/cargo-home/bin/cargo" "+209b2be09fcaff937480d1fbbe8b31646e361c7a" "generate-lockfile" "--manifest-path" "Cargo.toml" "-Zno-index-update"`
[INFO] [stderr]     Blocking waiting for file lock on package cache
[INFO] running `"/workspace/cargo-home/bin/cargo" "+209b2be09fcaff937480d1fbbe8b31646e361c7a" "fetch" "--locked" "--manifest-path" "Cargo.toml"`
[INFO] [stderr] warning: dependency (crawler) specified without providing a local path, Git repository, or version to use. This will be considered an error in future versions
[INFO] [stderr]     Blocking waiting for file lock on package cache
[INFO] [stderr]     Blocking waiting for file lock on package cache
[INFO] running `"docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-6/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-6/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "MAP_USER_ID=0" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--network" "none" "rustops/crates-build-env" "/opt/rustwide/cargo-home/bin/cargo" "+209b2be09fcaff937480d1fbbe8b31646e361c7a" "check" "--frozen" "--all" "--all-targets"`
[INFO] [stderr] WARNING: Your kernel does not support swap limit capabilities or the cgroup is not mounted. Memory limited without swap.
[INFO] [stdout] 0d50819677822b339ecf4f5544193131a99f6c6b94c68f4de67f671130cd21de
[INFO] running `"docker" "start" "-a" "0d50819677822b339ecf4f5544193131a99f6c6b94c68f4de67f671130cd21de"`
[INFO] [stderr] warning: dependency (crawler) specified without providing a local path, Git repository, or version to use. This will be considered an error in future versions
[INFO] [stderr]     Checking futures v0.1.29
[INFO] [stderr]    Compiling maybe-uninit v2.0.0
[INFO] [stderr]     Checking either v1.5.3
[INFO] [stderr]    Compiling syn v1.0.17
[INFO] [stderr]    Compiling serde v1.0.106
[INFO] [stderr]    Compiling failure_derive v0.1.7
[INFO] [stderr]     Checking quick-error v1.2.3
[INFO] [stderr]     Checking termcolor v1.1.0
[INFO] [stderr]     Checking crawler v0.1.0
[INFO] [stderr]    Compiling crossbeam-utils v0.7.2
[INFO] [stderr]    Compiling memoffset v0.5.4
[INFO] [stderr]    Compiling crossbeam-epoch v0.8.2
[INFO] [stderr]     Checking lock_api v0.3.4
[INFO] [stderr]     Checking idna v0.1.5
[INFO] [stderr]     Checking num_cpus v1.12.0
[INFO] [stderr]    Compiling parking_lot_core v0.6.2
[INFO] [stderr]    Compiling parking_lot v0.9.0
[INFO] [stderr]    Compiling hyper v0.12.35
[INFO] [stderr]     Checking backtrace-sys v0.1.35
[INFO] [stderr]     Checking mio v0.6.21
[INFO] [stderr]     Checking humantime v1.3.0
[INFO] [stderr]     Checking backtrace v0.3.46
[INFO] [stderr]     Checking bytes v0.4.12
[INFO] [stderr]     Checking env_logger v0.6.2
[INFO] [stderr]     Checking smallvec v0.6.13
[INFO] [stderr]     Checking url v1.7.2
[INFO] [stderr]     Checking http v0.1.21
[INFO] [stderr]     Checking string v0.2.1
[INFO] [stderr]     Checking crossbeam-queue v0.2.1
[INFO] [stderr]     Checking mio-uds v0.6.7
[INFO] [stderr]     Checking crossbeam-deque v0.7.3
[INFO] [stderr]     Checking tokio-executor v0.1.10
[INFO] [stderr]     Checking tokio-io v0.1.13
[INFO] [stderr]     Checking tokio-sync v0.1.8
[INFO] [stderr]     Checking tokio-buf v0.1.1
[INFO] [stderr]     Checking want v0.2.0
[INFO] [stderr]     Checking futures-cpupool v0.1.8
[INFO] [stderr]     Checking tokio-threadpool v0.1.18
[INFO] [stderr]     Checking tokio-current-thread v0.1.7
[INFO] [stderr]     Checking tokio-timer v0.2.13
[INFO] [stderr]     Checking tokio-codec v0.1.2
[INFO] [stderr]     Checking tokio-reactor v0.1.12
[INFO] [stderr]     Checking tokio-fs v0.1.7
[INFO] [stderr]     Checking tokio-tcp v0.1.4
[INFO] [stderr]     Checking tokio-uds v0.2.6
[INFO] [stderr]     Checking tokio-udp v0.1.6
[INFO] [stderr]     Checking http-body v0.1.0
[INFO] [stderr]     Checking h2 v0.1.26
[INFO] [stderr]     Checking tokio v0.1.22
[INFO] [stderr]    Compiling synstructure v0.12.3
[INFO] [stderr]    Compiling serde_derive v1.0.106
[INFO] [stderr]     Checking serde_json v1.0.51
[INFO] [stderr]     Checking hashbrown v0.4.0
[INFO] [stderr]     Checking failure v0.1.7
[INFO] [stderr]     Checking webcrawl v0.1.0 (/opt/rustwide/workdir)
[INFO] [stderr] error[E0432]: unresolved imports `crawler::prelude`, `serde_json`, `env_logger`
[INFO] [stderr]   --> src/main.rs:65:14
[INFO] [stderr]    |
[INFO] [stderr] 65 | use crawler::prelude::*;
[INFO] [stderr]    |              ^^^^^^^ could not find `prelude` in `crawler`
[INFO] [stderr] ...
[INFO] [stderr] 75 | use serde_json;
[INFO] [stderr]    |     ^^^^^^^^^^
[INFO] [stderr] ...
[INFO] [stderr] 86 | use env_logger;
[INFO] [stderr]    |     ^^^^^^^^^^
[INFO] [stderr] 
[INFO] [stderr] error[E0433]: failed to resolve: use of undeclared type or module `Crawler`
[INFO] [stderr]    --> src/main.rs:106:23
[INFO] [stderr]     |
[INFO] [stderr] 106 |         let crawler = Crawler::new(url, fetch, |base, html| parse(base, html, parse_url)).unwrap();
[INFO] [stderr]     |                       ^^^^^^^ use of undeclared type or module `Crawler`
[INFO] [stderr] 
[INFO] [stderr] error[E0412]: cannot find type `Urls` in this scope
[INFO] [stderr]    --> src/main.rs:95:14
[INFO] [stderr]     |
[INFO] [stderr] 95  |     Finished(Urls),
[INFO] [stderr]     |              ^^^^ help: a struct with a similar name exists: `Url`
[INFO] [stderr]     | 
[INFO] [stderr]    ::: /opt/rustwide/cargo-home/registry/src/github.com-1ecc6299db9ec823/url-1.7.2/src/lib.rs:154:1
[INFO] [stderr]     |
[INFO] [stderr] 154 | pub struct Url {
[INFO] [stderr]     | -------------- similarly named struct `Url` defined here
[INFO] [stderr] 
[INFO] [stderr] error[E0425]: cannot find value `fetch` in this scope
[INFO] [stderr]    --> src/main.rs:106:41
[INFO] [stderr]     |
[INFO] [stderr] 106 |         let crawler = Crawler::new(url, fetch, |base, html| parse(base, html, parse_url)).unwrap();
[INFO] [stderr]     |                                         ^^^^^ not found in this scope
[INFO] [stderr] 
[INFO] [stderr] error[E0425]: cannot find function `parse` in this scope
[INFO] [stderr]    --> src/main.rs:106:61
[INFO] [stderr]     |
[INFO] [stderr] 106 |         let crawler = Crawler::new(url, fetch, |base, html| parse(base, html, parse_url)).unwrap();
[INFO] [stderr]     |                                                             ^^^^^ not found in this scope
[INFO] [stderr]     |
[INFO] [stderr] help: possible candidate is found in another module, you can import it into scope
[INFO] [stderr]     |
[INFO] [stderr] 62  | use url::form_urlencoded::parse;
[INFO] [stderr]     |
[INFO] [stderr] 
[INFO] [stderr] error[E0425]: cannot find value `parse_url` in this scope
[INFO] [stderr]    --> src/main.rs:106:79
[INFO] [stderr]     |
[INFO] [stderr] 106 |         let crawler = Crawler::new(url, fetch, |base, html| parse(base, html, parse_url)).unwrap();
[INFO] [stderr]     |                                                                               ^^^^^^^^^ not found in this scope
[INFO] [stderr] 
[INFO] [stderr] warning: trait objects without an explicit `dyn` are deprecated
[INFO] [stderr]    --> src/main.rs:145:16
[INFO] [stderr]     |
[INFO] [stderr] 145 |         -> Box<Future<Item=Response<Body>, Error=hyper::Error> + Send> {
[INFO] [stderr]     |                ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ help: use `dyn`: `dyn Future<Item=Response<Body>, Error=hyper::Error> + Send`
[INFO] [stderr]     |
[INFO] [stderr]     = note: `#[warn(bare_trait_objects)]` on by default
[INFO] [stderr] 
[INFO] [stderr] error[E0432]: unresolved imports `crawler::prelude`, `serde_json`, `env_logger`
[INFO] [stderr]   --> src/main.rs:65:14
[INFO] [stderr]    |
[INFO] [stderr] 65 | use crawler::prelude::*;
[INFO] [stderr]    |              ^^^^^^^ could not find `prelude` in `crawler`
[INFO] [stderr] ...
[INFO] [stderr] 75 | use serde_json;
[INFO] [stderr]    |     ^^^^^^^^^^
[INFO] [stderr] ...
[INFO] [stderr] 86 | use env_logger;
[INFO] [stderr]    |     ^^^^^^^^^^
[INFO] [stderr] 
[INFO] [stderr] error[E0433]: failed to resolve: use of undeclared type or module `Crawler`
[INFO] [stderr]    --> src/main.rs:106:23
[INFO] [stderr]     |
[INFO] [stderr] 106 |         let crawler = Crawler::new(url, fetch, |base, html| parse(base, html, parse_url)).unwrap();
[INFO] [stderr]     |                       ^^^^^^^ use of undeclared type or module `Crawler`
[INFO] [stderr] 
[INFO] [stderr] error[E0412]: cannot find type `Urls` in this scope
[INFO] [stderr]    --> src/main.rs:95:14
[INFO] [stderr]     |
[INFO] [stderr] 95  |     Finished(Urls),
[INFO] [stderr]     |              ^^^^ help: a struct with a similar name exists: `Url`
[INFO] [stderr]     | 
[INFO] [stderr]    ::: /opt/rustwide/cargo-home/registry/src/github.com-1ecc6299db9ec823/url-1.7.2/src/lib.rs:154:1
[INFO] [stderr]     |
[INFO] [stderr] 154 | pub struct Url {
[INFO] [stderr]     | -------------- similarly named struct `Url` defined here
[INFO] [stderr] 
[INFO] [stderr] error[E0425]: cannot find value `fetch` in this scope
[INFO] [stderr]    --> src/main.rs:106:41
[INFO] [stderr]     |
[INFO] [stderr] 106 |         let crawler = Crawler::new(url, fetch, |base, html| parse(base, html, parse_url)).unwrap();
[INFO] [stderr]     |                                         ^^^^^ not found in this scope
[INFO] [stderr] 
[INFO] [stderr] error[E0425]: cannot find function `parse` in this scope
[INFO] [stderr]    --> src/main.rs:106:61
[INFO] [stderr]     |
[INFO] [stderr] 106 |         let crawler = Crawler::new(url, fetch, |base, html| parse(base, html, parse_url)).unwrap();
[INFO] [stderr]     |                                                             ^^^^^ not found in this scope
[INFO] [stderr]     |
[INFO] [stderr] help: possible candidate is found in another module, you can import it into scope
[INFO] [stderr]     |
[INFO] [stderr] 62  | use url::form_urlencoded::parse;
[INFO] [stderr]     |
[INFO] [stderr] 
[INFO] [stderr] error[E0425]: cannot find value `parse_url` in this scope
[INFO] [stderr]    --> src/main.rs:106:79
[INFO] [stderr]     |
[INFO] [stderr] 106 |         let crawler = Crawler::new(url, fetch, |base, html| parse(base, html, parse_url)).unwrap();
[INFO] [stderr]     |                                                                               ^^^^^^^^^ not found in this scope
[INFO] [stderr] 
[INFO] [stderr] warning: trait objects without an explicit `dyn` are deprecated
[INFO] [stderr]    --> src/main.rs:145:16
[INFO] [stderr]     |
[INFO] [stderr] 145 |         -> Box<Future<Item=Response<Body>, Error=hyper::Error> + Send> {
[INFO] [stderr]     |                ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ help: use `dyn`: `dyn Future<Item=Response<Body>, Error=hyper::Error> + Send`
[INFO] [stderr]     |
[INFO] [stderr]     = note: `#[warn(bare_trait_objects)]` on by default
[INFO] [stderr] 
[INFO] [stderr] error: aborting due to 6 previous errors
[INFO] [stderr] 
[INFO] [stderr] Some errors have detailed explanations: E0412, E0425, E0432, E0433.
[INFO] [stderr] For more information about an error, try `rustc --explain E0412`.
[INFO] [stderr] error: could not compile `webcrawl`.
[INFO] [stderr] 
[INFO] [stderr] To learn more, run the command again with --verbose.
[INFO] [stderr] warning: build failed, waiting for other jobs to finish...
[INFO] [stderr] error: aborting due to 6 previous errors
[INFO] [stderr] 
[INFO] [stderr] Some errors have detailed explanations: E0412, E0425, E0432, E0433.
[INFO] [stderr] For more information about an error, try `rustc --explain E0412`.
[INFO] [stderr] error: could not compile `webcrawl`.
[INFO] [stderr] 
[INFO] [stderr] To learn more, run the command again with --verbose.
[INFO] running `"docker" "inspect" "0d50819677822b339ecf4f5544193131a99f6c6b94c68f4de67f671130cd21de"`
[INFO] running `"docker" "rm" "-f" "0d50819677822b339ecf4f5544193131a99f6c6b94c68f4de67f671130cd21de"`
[INFO] [stdout] 0d50819677822b339ecf4f5544193131a99f6c6b94c68f4de67f671130cd21de
