[INFO] fetching crate bareun_rs 0.1.0... [INFO] checking bareun_rs-0.1.0 against master#8c32e313cccf7df531e2d49ffb8227bb92304aee for pr-145608-1 [INFO] extracting crate bareun_rs 0.1.0 into /workspace/builds/worker-5-tc1/source [INFO] started tweaking crates.io crate bareun_rs 0.1.0 [INFO] finished tweaking crates.io crate bareun_rs 0.1.0 [INFO] tweaked toml for crates.io crate bareun_rs 0.1.0 written to /workspace/builds/worker-5-tc1/source/Cargo.toml [INFO] validating manifest of crates.io crate bareun_rs 0.1.0 on toolchain 8c32e313cccf7df531e2d49ffb8227bb92304aee [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+8c32e313cccf7df531e2d49ffb8227bb92304aee" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }` [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+8c32e313cccf7df531e2d49ffb8227bb92304aee" "generate-lockfile" "--manifest-path" "Cargo.toml", kill_on_drop: false }` [INFO] [stderr] Updating crates.io index [INFO] [stderr] Locking 140 packages to latest compatible versions [INFO] [stderr] Adding prost v0.12.6 (available: v0.14.1) [INFO] [stderr] Adding tonic v0.11.0 (available: v0.14.2) [INFO] [stderr] Adding tonic-build v0.11.0 (available: v0.14.2) [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+8c32e313cccf7df531e2d49ffb8227bb92304aee" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }` [INFO] [stderr] Blocking waiting for file lock on package cache [INFO] [stderr] Blocking waiting for file lock on package cache [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:7ad1b28ee6f5f7f699f6cf7015098d6ccdd96d6f2d78dd06228f5b4c9faf309c" "/opt/rustwide/cargo-home/bin/cargo" "+8c32e313cccf7df531e2d49ffb8227bb92304aee" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }` [INFO] [stdout] 2507e3875014f0a386afed3509c714ab4a4afb902ff2a3ffccf295bf9c171ef5 [INFO] running `Command { std: "docker" "start" "-a" "2507e3875014f0a386afed3509c714ab4a4afb902ff2a3ffccf295bf9c171ef5", kill_on_drop: false }` [INFO] running `Command { std: "docker" "inspect" "2507e3875014f0a386afed3509c714ab4a4afb902ff2a3ffccf295bf9c171ef5", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "2507e3875014f0a386afed3509c714ab4a4afb902ff2a3ffccf295bf9c171ef5", kill_on_drop: false }` [INFO] [stdout] 2507e3875014f0a386afed3509c714ab4a4afb902ff2a3ffccf295bf9c171ef5 [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:7ad1b28ee6f5f7f699f6cf7015098d6ccdd96d6f2d78dd06228f5b4c9faf309c" "/opt/rustwide/cargo-home/bin/cargo" "+8c32e313cccf7df531e2d49ffb8227bb92304aee" "check" "--frozen" "--all" "--all-targets" "--message-format=json", kill_on_drop: false }` [INFO] [stdout] d29111a62f8409728a2d3dfe8f0c7d4827648d4308af0eee169290d7fbd3543e [INFO] running `Command { std: "docker" "start" "-a" "d29111a62f8409728a2d3dfe8f0c7d4827648d4308af0eee169290d7fbd3543e", kill_on_drop: false }` [INFO] [stderr] Checking smallvec v1.15.1 [INFO] [stderr] Compiling either v1.15.0 [INFO] [stderr] Compiling libc v0.2.175 [INFO] [stderr] Compiling rustix v1.0.8 [INFO] [stderr] Compiling getrandom v0.3.3 [INFO] [stderr] Compiling prettyplease v0.2.37 [INFO] [stderr] Compiling bytes v1.10.1 [INFO] [stderr] Compiling regex-syntax v0.8.6 [INFO] [stderr] Compiling linux-raw-sys v0.9.4 [INFO] [stderr] Compiling itertools v0.12.1 [INFO] [stderr] Compiling petgraph v0.6.5 [INFO] [stderr] Checking parking_lot_core v0.9.11 [INFO] [stderr] Checking parking_lot v0.12.4 [INFO] [stderr] Checking tokio v1.47.1 [INFO] [stderr] Compiling serde_derive v1.0.219 [INFO] [stderr] Checking axum-core v0.3.4 [INFO] [stderr] Compiling regex-automata v0.4.10 [INFO] [stderr] Compiling prost-derive v0.12.6 [INFO] [stderr] Compiling tempfile v3.21.0 [INFO] [stderr] Compiling prost v0.12.6 [INFO] [stderr] Compiling regex v1.11.2 [INFO] [stderr] Compiling prost-types v0.12.6 [INFO] [stderr] Checking serde v1.0.219 [INFO] [stderr] Compiling prost-build v0.12.6 [INFO] [stderr] Checking tokio-util v0.7.16 [INFO] [stderr] Checking tokio-io-timeout v1.2.1 [INFO] [stderr] Checking tokio-stream v0.1.17 [INFO] [stderr] Compiling tonic-build v0.11.0 [INFO] [stderr] Checking h2 v0.3.27 [INFO] [stderr] Checking tower v0.4.13 [INFO] [stderr] Checking serde_json v1.0.143 [INFO] [stderr] Compiling bareun_rs v0.1.0 (/opt/rustwide/workdir) [INFO] [stderr] Checking hyper v0.14.32 [INFO] [stderr] Checking hyper-timeout v0.4.1 [INFO] [stderr] Checking axum v0.6.20 [INFO] [stderr] Checking tonic v0.11.0 [INFO] [stdout] warning: unused import: `Error` [INFO] [stdout] --> src/lang_service_client.rs:3:42 [INFO] [stdout] | [INFO] [stdout] 3 | use tonic::transport::{channel, Channel, Error}; [INFO] [stdout] | ^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `Tag` [INFO] [stdout] --> src/tagger.rs:3:43 [INFO] [stdout] | [INFO] [stdout] 3 | use crate::bareun::morpheme::{OutOfVocab, Tag}; [INFO] [stdout] | ^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `MAX_MESSAGE_LENGTH` [INFO] [stdout] --> src/tagger.rs:7:63 [INFO] [stdout] | [INFO] [stdout] 7 | use crate::lang_service_client::{BareunLanguageServiceClient, MAX_MESSAGE_LENGTH}; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `tonic::transport::Endpoint` [INFO] [stdout] --> src/tagger.rs:11:5 [INFO] [stdout] | [INFO] [stdout] 11 | use tonic::transport::Endpoint; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `std::collections::HashMap` [INFO] [stdout] --> src/tokenizer.rs:1:5 [INFO] [stdout] | [INFO] [stdout] 1 | use std::collections::HashMap; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused imports: `SegmentToken` and `TextSpan` [INFO] [stdout] --> src/tokenizer.rs:3:47 [INFO] [stdout] | [INFO] [stdout] 3 | use crate::bareun::{Segment, SegmentSentence, SegmentToken, TextSpan, TokenizeResponse}; [INFO] [stdout] | ^^^^^^^^^^^^ ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `MAX_MESSAGE_LENGTH` [INFO] [stdout] --> src/tokenizer.rs:4:63 [INFO] [stdout] | [INFO] [stdout] 4 | use crate::lang_service_client::{BareunLanguageServiceClient, MAX_MESSAGE_LENGTH}; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused imports: `Deserialize` and `Serialize` [INFO] [stdout] --> src/tokenizer.rs:5:13 [INFO] [stdout] | [INFO] [stdout] 5 | use serde::{Deserialize, Serialize}; [INFO] [stdout] | ^^^^^^^^^^^ ^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `serde_json::json` [INFO] [stdout] --> src/tokenizer.rs:7:5 [INFO] [stdout] | [INFO] [stdout] 7 | use serde_json::json; [INFO] [stdout] | ^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `tonic::transport::Endpoint` [INFO] [stdout] --> src/tokenizer.rs:8:5 [INFO] [stdout] | [INFO] [stdout] 8 | use tonic::transport::Endpoint; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `Error` [INFO] [stdout] --> src/lang_service_client.rs:3:42 [INFO] [stdout] | [INFO] [stdout] 3 | use tonic::transport::{channel, Channel, Error}; [INFO] [stdout] | ^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `Tag` [INFO] [stdout] --> src/tagger.rs:3:43 [INFO] [stdout] | [INFO] [stdout] 3 | use crate::bareun::morpheme::{OutOfVocab, Tag}; [INFO] [stdout] | ^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `MAX_MESSAGE_LENGTH` [INFO] [stdout] --> src/tagger.rs:7:63 [INFO] [stdout] | [INFO] [stdout] 7 | use crate::lang_service_client::{BareunLanguageServiceClient, MAX_MESSAGE_LENGTH}; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `tonic::transport::Endpoint` [INFO] [stdout] --> src/tagger.rs:11:5 [INFO] [stdout] | [INFO] [stdout] 11 | use tonic::transport::Endpoint; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `std::collections::HashMap` [INFO] [stdout] --> src/tokenizer.rs:1:5 [INFO] [stdout] | [INFO] [stdout] 1 | use std::collections::HashMap; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused imports: `SegmentToken` and `TextSpan` [INFO] [stdout] --> src/tokenizer.rs:3:47 [INFO] [stdout] | [INFO] [stdout] 3 | use crate::bareun::{Segment, SegmentSentence, SegmentToken, TextSpan, TokenizeResponse}; [INFO] [stdout] | ^^^^^^^^^^^^ ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `MAX_MESSAGE_LENGTH` [INFO] [stdout] --> src/tokenizer.rs:4:63 [INFO] [stdout] | [INFO] [stdout] 4 | use crate::lang_service_client::{BareunLanguageServiceClient, MAX_MESSAGE_LENGTH}; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused imports: `Deserialize` and `Serialize` [INFO] [stdout] --> src/tokenizer.rs:5:13 [INFO] [stdout] | [INFO] [stdout] 5 | use serde::{Deserialize, Serialize}; [INFO] [stdout] | ^^^^^^^^^^^ ^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `serde_json::json` [INFO] [stdout] --> src/tokenizer.rs:7:5 [INFO] [stdout] | [INFO] [stdout] 7 | use serde_json::json; [INFO] [stdout] | ^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `tonic::transport::Endpoint` [INFO] [stdout] --> src/tokenizer.rs:8:5 [INFO] [stdout] | [INFO] [stdout] 8 | use tonic::transport::Endpoint; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `pb_map_to_set` is never used [INFO] [stdout] --> src/custom_dict.rs:39:8 [INFO] [stdout] | [INFO] [stdout] 39 | pub fn pb_map_to_set(ds: &DictSet) -> HashSet { [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(dead_code)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: constant `MAX_MESSAGE_LENGTH` is never used [INFO] [stdout] --> src/lang_service_client.rs:12:11 [INFO] [stdout] | [INFO] [stdout] 12 | pub const MAX_MESSAGE_LENGTH: usize = 100 * 1024 * 1024; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `metadata` is never read [INFO] [stdout] --> src/lang_service_client.rs:19:5 [INFO] [stdout] | [INFO] [stdout] 13 | pub struct BareunLanguageServiceClient { [INFO] [stdout] | --------------------------- field in this struct [INFO] [stdout] ... [INFO] [stdout] 19 | metadata: HashMap, [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `phrase` is never read [INFO] [stdout] --> src/tagger.rs:17:5 [INFO] [stdout] | [INFO] [stdout] 12 | pub struct Tagged { [INFO] [stdout] | ------ field in this struct [INFO] [stdout] ... [INFO] [stdout] 17 | phrase: String, [INFO] [stdout] | ^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `phrase` is never read [INFO] [stdout] --> src/tokenizer.rs:18:5 [INFO] [stdout] | [INFO] [stdout] 13 | pub struct Tokenized { [INFO] [stdout] | --------- field in this struct [INFO] [stdout] ... [INFO] [stdout] 18 | phrase: String, [INFO] [stdout] | ^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `pb_map_to_set` is never used [INFO] [stdout] --> src/custom_dict.rs:39:8 [INFO] [stdout] | [INFO] [stdout] 39 | pub fn pb_map_to_set(ds: &DictSet) -> HashSet { [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(dead_code)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: constant `MAX_MESSAGE_LENGTH` is never used [INFO] [stdout] --> src/lang_service_client.rs:12:11 [INFO] [stdout] | [INFO] [stdout] 12 | pub const MAX_MESSAGE_LENGTH: usize = 100 * 1024 * 1024; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `metadata` is never read [INFO] [stdout] --> src/lang_service_client.rs:19:5 [INFO] [stdout] | [INFO] [stdout] 13 | pub struct BareunLanguageServiceClient { [INFO] [stdout] | --------------------------- field in this struct [INFO] [stdout] ... [INFO] [stdout] 19 | metadata: HashMap, [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `phrase` is never read [INFO] [stdout] --> src/tagger.rs:17:5 [INFO] [stdout] | [INFO] [stdout] 12 | pub struct Tagged { [INFO] [stdout] | ------ field in this struct [INFO] [stdout] ... [INFO] [stdout] 17 | phrase: String, [INFO] [stdout] | ^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `phrase` is never read [INFO] [stdout] --> src/tokenizer.rs:18:5 [INFO] [stdout] | [INFO] [stdout] 13 | pub struct Tokenized { [INFO] [stdout] | --------- field in this struct [INFO] [stdout] ... [INFO] [stdout] 18 | phrase: String, [INFO] [stdout] | ^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `bareun_rs::Tokenized` [INFO] [stdout] --> tests/test_tokenizer.rs:3:9 [INFO] [stdout] | [INFO] [stdout] 3 | use bareun_rs::Tokenized; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this method takes 4 arguments but 2 arguments were supplied [INFO] [stdout] --> tests/test_tokenizer.rs:16:29 [INFO] [stdout] | [INFO] [stdout] 16 | tokenizer.await.seg(TEST_STR, false), [INFO] [stdout] | ^^^----------------- two arguments of type `bool` and `bool` are missing [INFO] [stdout] | [INFO] [stdout] note: method defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:381:18 [INFO] [stdout] | [INFO] [stdout] 381 | pub async fn seg( [INFO] [stdout] | ^^^ [INFO] [stdout] help: provide the arguments [INFO] [stdout] | [INFO] [stdout] 16 | tokenizer.await.seg(TEST_STR, false, /* bool */, /* bool */), [INFO] [stdout] | ++++++++++++++++++++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0369]: binary operation `==` cannot be applied to type `impl Future` [INFO] [stdout] --> tests/test_tokenizer.rs:15:9 [INFO] [stdout] | [INFO] [stdout] 15 | / assert_eq!( [INFO] [stdout] 16 | | tokenizer.await.seg(TEST_STR, false), [INFO] [stdout] 17 | | vec![ [INFO] [stdout] 18 | | vec!["오늘".to_string(), "은".to_string()], [INFO] [stdout] ... | [INFO] [stdout] 29 | | ); [INFO] [stdout] | | ^ [INFO] [stdout] | | | [INFO] [stdout] | |_________impl Future [INFO] [stdout] | Vec> [INFO] [stdout] | [INFO] [stdout] = note: this error originates in the macro `assert_eq` (in Nightly builds, run with -Z macro-backtrace for more info) [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0277]: `impl Future` doesn't implement `Debug` [INFO] [stdout] --> tests/test_tokenizer.rs:15:9 [INFO] [stdout] | [INFO] [stdout] 15 | / assert_eq!( [INFO] [stdout] 16 | | tokenizer.await.seg(TEST_STR, false), [INFO] [stdout] 17 | | vec![ [INFO] [stdout] 18 | | vec!["오늘".to_string(), "은".to_string()], [INFO] [stdout] ... | [INFO] [stdout] 29 | | ); [INFO] [stdout] | |_________^ the trait `Debug` is not implemented for `impl Future` [INFO] [stdout] | [INFO] [stdout] = note: this error originates in the macro `assert_eq` (in Nightly builds, run with -Z macro-backtrace for more info) [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:34:93 [INFO] [stdout] | [INFO] [stdout] 34 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 34 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `seg` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:36:23 [INFO] [stdout] | [INFO] [stdout] 36 | tokenizer.seg(TEST_STR, true, true), [INFO] [stdout] | ^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 36 | tokenizer.await.seg(TEST_STR, true, true), [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:54:93 [INFO] [stdout] | [INFO] [stdout] 54 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 54 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `seg` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:55:30 [INFO] [stdout] | [INFO] [stdout] 55 | let temp = tokenizer.seg(TEST_STR, true); [INFO] [stdout] | ^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 55 | let temp = tokenizer.await.seg(TEST_STR, true); [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:76:93 [INFO] [stdout] | [INFO] [stdout] 76 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 76 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `seg` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:78:23 [INFO] [stdout] | [INFO] [stdout] 78 | tokenizer.seg(TEST_STR), [INFO] [stdout] | ^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 78 | tokenizer.await.seg(TEST_STR), [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:96:93 [INFO] [stdout] | [INFO] [stdout] 96 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 96 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `nouns` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:98:23 [INFO] [stdout] | [INFO] [stdout] 98 | tokenizer.nouns(TEST_STR), [INFO] [stdout] | ^^^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 98 | tokenizer.await.nouns(TEST_STR), [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:105:93 [INFO] [stdout] | [INFO] [stdout] 105 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 105 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `tokenize` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:106:35 [INFO] [stdout] | [INFO] [stdout] 106 | let tokenized = tokenizer.tokenize(TEST_STR); [INFO] [stdout] | ^^^^^^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 106 | let tokenized = tokenizer.await.tokenize(TEST_STR); [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:115:93 [INFO] [stdout] | [INFO] [stdout] 115 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 115 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `tokenize` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:116:35 [INFO] [stdout] | [INFO] [stdout] 116 | let tokenized = tokenizer.tokenize(TEST_STR); [INFO] [stdout] | ^^^^^^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 116 | let tokenized = tokenizer.await.tokenize(TEST_STR); [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:125:93 [INFO] [stdout] | [INFO] [stdout] 125 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 125 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `tokenize` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:126:35 [INFO] [stdout] | [INFO] [stdout] 126 | let tokenized = tokenizer.tokenize(TEST_STR); [INFO] [stdout] | ^^^^^^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 126 | let tokenized = tokenizer.await.tokenize(TEST_STR); [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:132:93 [INFO] [stdout] | [INFO] [stdout] 132 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 132 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `tokenize` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:133:35 [INFO] [stdout] | [INFO] [stdout] 133 | let tokenized = tokenizer.tokenize(TEST_STR); [INFO] [stdout] | ^^^^^^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 133 | let tokenized = tokenizer.await.tokenize(TEST_STR); [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:139:93 [INFO] [stdout] | [INFO] [stdout] 139 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 139 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `tokenize` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:140:35 [INFO] [stdout] | [INFO] [stdout] 140 | let tokenized = tokenizer.tokenize(TEST_STR); [INFO] [stdout] | ^^^^^^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 140 | let tokenized = tokenizer.await.tokenize(TEST_STR); [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:149:93 [INFO] [stdout] | [INFO] [stdout] 149 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 149 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `tokenize` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:150:35 [INFO] [stdout] | [INFO] [stdout] 150 | let tokenized = tokenizer.tokenize(TEST_STR); [INFO] [stdout] | ^^^^^^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 150 | let tokenized = tokenizer.await.tokenize(TEST_STR); [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] Some errors have detailed explanations: E0061, E0277, E0308, E0369, E0599. [INFO] [stdout] [INFO] [stdout] For more information about an error, try `rustc --explain E0061`. [INFO] [stdout] [INFO] [stderr] error: could not compile `bareun_rs` (test "test_tokenizer") due to 23 previous errors; 1 warning emitted [INFO] running `Command { std: "docker" "inspect" "d29111a62f8409728a2d3dfe8f0c7d4827648d4308af0eee169290d7fbd3543e", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "d29111a62f8409728a2d3dfe8f0c7d4827648d4308af0eee169290d7fbd3543e", kill_on_drop: false }` [INFO] [stdout] d29111a62f8409728a2d3dfe8f0c7d4827648d4308af0eee169290d7fbd3543e