[INFO] fetching crate bareun_rs 0.1.0...
[INFO] testing bareun_rs-0.1.0 against master#c90bcb9571b7aab0d8beaa2ce8a998ffaf079d38 for pr-146098-7
[INFO] extracting crate bareun_rs 0.1.0 into /workspace/builds/worker-7-tc1/source
[INFO] started tweaking crates.io crate bareun_rs 0.1.0
[INFO] finished tweaking crates.io crate bareun_rs 0.1.0
[INFO] tweaked toml for crates.io crate bareun_rs 0.1.0 written to /workspace/builds/worker-7-tc1/source/Cargo.toml
[INFO] validating manifest of crates.io crate bareun_rs 0.1.0 on toolchain c90bcb9571b7aab0d8beaa2ce8a998ffaf079d38
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+c90bcb9571b7aab0d8beaa2ce8a998ffaf079d38" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }`
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+c90bcb9571b7aab0d8beaa2ce8a998ffaf079d38" "generate-lockfile" "--manifest-path" "Cargo.toml", kill_on_drop: false }`
[INFO] [stderr]     Updating crates.io index
[INFO] [stderr]      Locking 133 packages to latest compatible versions
[INFO] [stderr]       Adding prost v0.12.6 (available: v0.14.1)
[INFO] [stderr]       Adding tonic v0.11.0 (available: v0.14.2)
[INFO] [stderr]       Adding tonic-build v0.11.0 (available: v0.14.2)
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+c90bcb9571b7aab0d8beaa2ce8a998ffaf079d38" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }`
[INFO] [stderr]  Downloading crates ...
[INFO] [stderr]   Downloaded tokio-io-timeout v1.2.1
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-7-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-7-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:4848fb76d95f26979359cc7e45710b1dbc8f3acb7aeedee7c460d7702230f228" "/opt/rustwide/cargo-home/bin/cargo" "+c90bcb9571b7aab0d8beaa2ce8a998ffaf079d38" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }`
[INFO] [stdout] b657032cf89e2c99657b1631a1c5a10ef60e332fa2797fd966d3162d3e75ed6b
[INFO] running `Command { std: "docker" "start" "-a" "b657032cf89e2c99657b1631a1c5a10ef60e332fa2797fd966d3162d3e75ed6b", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "inspect" "b657032cf89e2c99657b1631a1c5a10ef60e332fa2797fd966d3162d3e75ed6b", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "b657032cf89e2c99657b1631a1c5a10ef60e332fa2797fd966d3162d3e75ed6b", kill_on_drop: false }`
[INFO] [stdout] b657032cf89e2c99657b1631a1c5a10ef60e332fa2797fd966d3162d3e75ed6b
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-7-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-7-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:4848fb76d95f26979359cc7e45710b1dbc8f3acb7aeedee7c460d7702230f228" "/opt/rustwide/cargo-home/bin/cargo" "+c90bcb9571b7aab0d8beaa2ce8a998ffaf079d38" "build" "--frozen" "--message-format=json", kill_on_drop: false }`
[INFO] [stdout] 3bbaa5a3216e6df066ca7e92d2b209cc94cdae30f3e8c312a7603889322eeba2
[INFO] running `Command { std: "docker" "start" "-a" "3bbaa5a3216e6df066ca7e92d2b209cc94cdae30f3e8c312a7603889322eeba2", kill_on_drop: false }`
[INFO] [stderr]    Compiling syn v2.0.110
[INFO] [stderr]    Compiling libc v0.2.177
[INFO] [stderr]    Compiling parking_lot_core v0.9.12
[INFO] [stderr]    Compiling smallvec v1.15.1
[INFO] [stderr]    Compiling lock_api v0.4.14
[INFO] [stderr]    Compiling hashbrown v0.16.0
[INFO] [stderr]    Compiling anyhow v1.0.100
[INFO] [stderr]    Compiling either v1.15.0
[INFO] [stderr]    Compiling futures-task v0.3.31
[INFO] [stderr]    Compiling itertools v0.12.1
[INFO] [stderr]    Compiling rustix v1.1.2
[INFO] [stderr]    Compiling getrandom v0.3.4
[INFO] [stderr]    Compiling futures-util v0.3.31
[INFO] [stderr]    Compiling prettyplease v0.2.37
[INFO] [stderr]    Compiling regex-syntax v0.8.8
[INFO] [stderr]    Compiling linux-raw-sys v0.11.0
[INFO] [stderr]    Compiling indexmap v2.12.0
[INFO] [stderr]    Compiling bitflags v2.10.0
[INFO] [stderr]    Compiling signal-hook-registry v1.4.6
[INFO] [stderr]    Compiling socket2 v0.6.1
[INFO] [stderr]    Compiling mio v1.1.0
[INFO] [stderr]    Compiling parking_lot v0.12.5
[INFO] [stderr]    Compiling getrandom v0.2.16
[INFO] [stderr]    Compiling rand_core v0.6.4
[INFO] [stderr]    Compiling indexmap v1.9.3
[INFO] [stderr]    Compiling fixedbitset v0.4.2
[INFO] [stderr]    Compiling fastrand v2.3.0
[INFO] [stderr]    Compiling rand_chacha v0.3.1
[INFO] [stderr]    Compiling socket2 v0.5.10
[INFO] [stderr]    Compiling axum-core v0.3.4
[INFO] [stderr]    Compiling futures-channel v0.3.31
[INFO] [stderr]    Compiling regex-automata v0.4.13
[INFO] [stderr]    Compiling petgraph v0.6.5
[INFO] [stderr]    Compiling log v0.4.28
[INFO] [stderr]    Compiling hashbrown v0.12.3
[INFO] [stderr]    Compiling rand v0.8.5
[INFO] [stderr]    Compiling tempfile v3.23.0
[INFO] [stderr]    Compiling axum v0.6.20
[INFO] [stderr]    Compiling percent-encoding v2.3.2
[INFO] [stderr]    Compiling sync_wrapper v0.1.2
[INFO] [stderr]    Compiling matchit v0.7.3
[INFO] [stderr]    Compiling serde_json v1.0.145
[INFO] [stderr]    Compiling regex v1.12.2
[INFO] [stderr]    Compiling tokio-macros v2.6.0
[INFO] [stderr]    Compiling prost-derive v0.12.6
[INFO] [stderr]    Compiling tracing-attributes v0.1.30
[INFO] [stderr]    Compiling pin-project-internal v1.1.10
[INFO] [stderr]    Compiling serde_derive v1.0.228
[INFO] [stderr]    Compiling async-trait v0.1.89
[INFO] [stderr]    Compiling async-stream-impl v0.3.6
[INFO] [stderr]    Compiling tokio v1.48.0
[INFO] [stderr]    Compiling async-stream v0.3.6
[INFO] [stderr]    Compiling tracing v0.1.41
[INFO] [stderr]    Compiling prost v0.12.6
[INFO] [stderr]    Compiling pin-project v1.1.10
[INFO] [stderr]    Compiling prost-types v0.12.6
[INFO] [stderr]    Compiling prost-build v0.12.6
[INFO] [stderr]    Compiling serde v1.0.228
[INFO] [stderr]    Compiling tonic-build v0.11.0
[INFO] [stderr]    Compiling bareun_rs v0.1.0 (/opt/rustwide/workdir)
[INFO] [stderr]    Compiling tokio-util v0.7.17
[INFO] [stderr]    Compiling tokio-io-timeout v1.2.1
[INFO] [stderr]    Compiling tokio-stream v0.1.17
[INFO] [stderr]    Compiling h2 v0.3.27
[INFO] [stderr]    Compiling tower v0.4.13
[INFO] [stderr]    Compiling hyper v0.14.32
[INFO] [stderr]    Compiling hyper-timeout v0.4.1
[INFO] [stderr]    Compiling tonic v0.11.0
[INFO] [stdout] warning: unused import: `Error`
[INFO] [stdout]  --> src/lang_service_client.rs:3:42
[INFO] [stdout]   |
[INFO] [stdout] 3 | use tonic::transport::{channel, Channel, Error};
[INFO] [stdout]   |                                          ^^^^^
[INFO] [stdout]   |
[INFO] [stdout]   = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `Tag`
[INFO] [stdout]  --> src/tagger.rs:3:43
[INFO] [stdout]   |
[INFO] [stdout] 3 | use crate::bareun::morpheme::{OutOfVocab, Tag};
[INFO] [stdout]   |                                           ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `MAX_MESSAGE_LENGTH`
[INFO] [stdout]  --> src/tagger.rs:7:63
[INFO] [stdout]   |
[INFO] [stdout] 7 | use crate::lang_service_client::{BareunLanguageServiceClient, MAX_MESSAGE_LENGTH};
[INFO] [stdout]   |                                                               ^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `tonic::transport::Endpoint`
[INFO] [stdout]   --> src/tagger.rs:11:5
[INFO] [stdout]    |
[INFO] [stdout] 11 | use tonic::transport::Endpoint;
[INFO] [stdout]    |     ^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `std::collections::HashMap`
[INFO] [stdout]  --> src/tokenizer.rs:1:5
[INFO] [stdout]   |
[INFO] [stdout] 1 | use std::collections::HashMap;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused imports: `SegmentToken` and `TextSpan`
[INFO] [stdout]  --> src/tokenizer.rs:3:47
[INFO] [stdout]   |
[INFO] [stdout] 3 | use crate::bareun::{Segment, SegmentSentence, SegmentToken, TextSpan, TokenizeResponse};
[INFO] [stdout]   |                                               ^^^^^^^^^^^^  ^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `MAX_MESSAGE_LENGTH`
[INFO] [stdout]  --> src/tokenizer.rs:4:63
[INFO] [stdout]   |
[INFO] [stdout] 4 | use crate::lang_service_client::{BareunLanguageServiceClient, MAX_MESSAGE_LENGTH};
[INFO] [stdout]   |                                                               ^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused imports: `Deserialize` and `Serialize`
[INFO] [stdout]  --> src/tokenizer.rs:5:13
[INFO] [stdout]   |
[INFO] [stdout] 5 | use serde::{Deserialize, Serialize};
[INFO] [stdout]   |             ^^^^^^^^^^^  ^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `serde_json::json`
[INFO] [stdout]  --> src/tokenizer.rs:7:5
[INFO] [stdout]   |
[INFO] [stdout] 7 | use serde_json::json;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `tonic::transport::Endpoint`
[INFO] [stdout]  --> src/tokenizer.rs:8:5
[INFO] [stdout]   |
[INFO] [stdout] 8 | use tonic::transport::Endpoint;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `pb_map_to_set` is never used
[INFO] [stdout]   --> src/custom_dict.rs:39:8
[INFO] [stdout]    |
[INFO] [stdout] 39 | pub fn pb_map_to_set(ds: &DictSet) -> HashSet<String> {
[INFO] [stdout]    |        ^^^^^^^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: constant `MAX_MESSAGE_LENGTH` is never used
[INFO] [stdout]   --> src/lang_service_client.rs:12:11
[INFO] [stdout]    |
[INFO] [stdout] 12 | pub const MAX_MESSAGE_LENGTH: usize = 100 * 1024 * 1024;
[INFO] [stdout]    |           ^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: field `metadata` is never read
[INFO] [stdout]   --> src/lang_service_client.rs:19:5
[INFO] [stdout]    |
[INFO] [stdout] 13 | pub struct BareunLanguageServiceClient {
[INFO] [stdout]    |            --------------------------- field in this struct
[INFO] [stdout] ...
[INFO] [stdout] 19 |     metadata: HashMap<String, String>,
[INFO] [stdout]    |     ^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: field `phrase` is never read
[INFO] [stdout]   --> src/tagger.rs:17:5
[INFO] [stdout]    |
[INFO] [stdout] 12 | pub struct Tagged {
[INFO] [stdout]    |            ------ field in this struct
[INFO] [stdout] ...
[INFO] [stdout] 17 |     phrase: String,
[INFO] [stdout]    |     ^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: field `phrase` is never read
[INFO] [stdout]   --> src/tokenizer.rs:18:5
[INFO] [stdout]    |
[INFO] [stdout] 13 | pub struct Tokenized {
[INFO] [stdout]    |            --------- field in this struct
[INFO] [stdout] ...
[INFO] [stdout] 18 |     phrase: String,
[INFO] [stdout]    |     ^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stderr]     Finished `dev` profile [unoptimized + debuginfo] target(s) in 1m 23s
[INFO] running `Command { std: "docker" "inspect" "3bbaa5a3216e6df066ca7e92d2b209cc94cdae30f3e8c312a7603889322eeba2", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "3bbaa5a3216e6df066ca7e92d2b209cc94cdae30f3e8c312a7603889322eeba2", kill_on_drop: false }`
[INFO] [stdout] 3bbaa5a3216e6df066ca7e92d2b209cc94cdae30f3e8c312a7603889322eeba2
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-7-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-7-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:4848fb76d95f26979359cc7e45710b1dbc8f3acb7aeedee7c460d7702230f228" "/opt/rustwide/cargo-home/bin/cargo" "+c90bcb9571b7aab0d8beaa2ce8a998ffaf079d38" "test" "--frozen" "--no-run" "--message-format=json", kill_on_drop: false }`
[INFO] [stdout] 0eeefc8044c2d76ffa371c7b39f98412d9e9975d99799c6020f95b088566d518
[INFO] running `Command { std: "docker" "start" "-a" "0eeefc8044c2d76ffa371c7b39f98412d9e9975d99799c6020f95b088566d518", kill_on_drop: false }`
[INFO] [stdout] warning: unused import: `Error`
[INFO] [stdout]  --> src/lang_service_client.rs:3:42
[INFO] [stdout]   |
[INFO] [stdout] 3 | use tonic::transport::{channel, Channel, Error};
[INFO] [stdout]   |                                          ^^^^^
[INFO] [stdout]   |
[INFO] [stdout]   = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `Tag`
[INFO] [stdout]  --> src/tagger.rs:3:43
[INFO] [stdout]   |
[INFO] [stdout] 3 | use crate::bareun::morpheme::{OutOfVocab, Tag};
[INFO] [stdout]   |                                           ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `MAX_MESSAGE_LENGTH`
[INFO] [stdout]  --> src/tagger.rs:7:63
[INFO] [stdout]   |
[INFO] [stdout] 7 | use crate::lang_service_client::{BareunLanguageServiceClient, MAX_MESSAGE_LENGTH};
[INFO] [stdout]   |                                                               ^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `tonic::transport::Endpoint`
[INFO] [stdout]   --> src/tagger.rs:11:5
[INFO] [stdout]    |
[INFO] [stdout] 11 | use tonic::transport::Endpoint;
[INFO] [stdout]    |     ^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `std::collections::HashMap`
[INFO] [stdout]  --> src/tokenizer.rs:1:5
[INFO] [stdout]   |
[INFO] [stdout] 1 | use std::collections::HashMap;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused imports: `SegmentToken` and `TextSpan`
[INFO] [stdout]  --> src/tokenizer.rs:3:47
[INFO] [stdout]   |
[INFO] [stdout] 3 | use crate::bareun::{Segment, SegmentSentence, SegmentToken, TextSpan, TokenizeResponse};
[INFO] [stdout]   |                                               ^^^^^^^^^^^^  ^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `MAX_MESSAGE_LENGTH`
[INFO] [stdout]  --> src/tokenizer.rs:4:63
[INFO] [stdout]   |
[INFO] [stdout] 4 | use crate::lang_service_client::{BareunLanguageServiceClient, MAX_MESSAGE_LENGTH};
[INFO] [stdout]   |                                                               ^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused imports: `Deserialize` and `Serialize`
[INFO] [stdout]  --> src/tokenizer.rs:5:13
[INFO] [stdout]   |
[INFO] [stdout] 5 | use serde::{Deserialize, Serialize};
[INFO] [stdout]   |             ^^^^^^^^^^^  ^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `serde_json::json`
[INFO] [stdout]  --> src/tokenizer.rs:7:5
[INFO] [stdout]   |
[INFO] [stdout] 7 | use serde_json::json;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `tonic::transport::Endpoint`
[INFO] [stdout]  --> src/tokenizer.rs:8:5
[INFO] [stdout]   |
[INFO] [stdout] 8 | use tonic::transport::Endpoint;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: function `pb_map_to_set` is never used
[INFO] [stdout]   --> src/custom_dict.rs:39:8
[INFO] [stdout]    |
[INFO] [stdout] 39 | pub fn pb_map_to_set(ds: &DictSet) -> HashSet<String> {
[INFO] [stdout]    |        ^^^^^^^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: constant `MAX_MESSAGE_LENGTH` is never used
[INFO] [stdout]   --> src/lang_service_client.rs:12:11
[INFO] [stdout]    |
[INFO] [stdout] 12 | pub const MAX_MESSAGE_LENGTH: usize = 100 * 1024 * 1024;
[INFO] [stdout]    |           ^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: field `metadata` is never read
[INFO] [stdout]   --> src/lang_service_client.rs:19:5
[INFO] [stdout]    |
[INFO] [stdout] 13 | pub struct BareunLanguageServiceClient {
[INFO] [stdout]    |            --------------------------- field in this struct
[INFO] [stdout] ...
[INFO] [stdout] 19 |     metadata: HashMap<String, String>,
[INFO] [stdout]    |     ^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: field `phrase` is never read
[INFO] [stdout]   --> src/tagger.rs:17:5
[INFO] [stdout]    |
[INFO] [stdout] 12 | pub struct Tagged {
[INFO] [stdout]    |            ------ field in this struct
[INFO] [stdout] ...
[INFO] [stdout] 17 |     phrase: String,
[INFO] [stdout]    |     ^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: field `phrase` is never read
[INFO] [stdout]   --> src/tokenizer.rs:18:5
[INFO] [stdout]    |
[INFO] [stdout] 13 | pub struct Tokenized {
[INFO] [stdout]    |            --------- field in this struct
[INFO] [stdout] ...
[INFO] [stdout] 18 |     phrase: String,
[INFO] [stdout]    |     ^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stderr]    Compiling bareun_rs v0.1.0 (/opt/rustwide/workdir)
[INFO] [stdout] warning: unused import: `bareun_rs::Tokenized`
[INFO] [stdout]  --> tests/test_tokenizer.rs:3:9
[INFO] [stdout]   |
[INFO] [stdout] 3 |     use bareun_rs::Tokenized;
[INFO] [stdout]   |         ^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout]   |
[INFO] [stdout]   = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0061]: this method takes 4 arguments but 2 arguments were supplied
[INFO] [stdout]    --> tests/test_tokenizer.rs:16:29
[INFO] [stdout]     |
[INFO] [stdout]  16 |             tokenizer.await.seg(TEST_STR, false),
[INFO] [stdout]     |                             ^^^----------------- two arguments of type `bool` and `bool` are missing
[INFO] [stdout]     |
[INFO] [stdout] note: method defined here
[INFO] [stdout]    --> /opt/rustwide/workdir/src/tokenizer.rs:381:18
[INFO] [stdout]     |
[INFO] [stdout] 381 |     pub async fn seg(
[INFO] [stdout]     |                  ^^^
[INFO] [stdout] help: provide the arguments
[INFO] [stdout]     |
[INFO] [stdout]  16 |             tokenizer.await.seg(TEST_STR, false, /* bool */, /* bool */),
[INFO] [stdout]     |                                                ++++++++++++++++++++++++
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0369]: binary operation `==` cannot be applied to type `impl Future<Output = bareun_rs::tokenizer::SegResult>`
[INFO] [stdout]   --> tests/test_tokenizer.rs:15:9
[INFO] [stdout]    |
[INFO] [stdout] 15 | /         assert_eq!(
[INFO] [stdout] 16 | |             tokenizer.await.seg(TEST_STR, false),
[INFO] [stdout] 17 | |             vec![
[INFO] [stdout] 18 | |                 vec!["오늘".to_string(), "은".to_string()],
[INFO] [stdout] ...  |
[INFO] [stdout] 29 | |         );
[INFO] [stdout]    | |         ^
[INFO] [stdout]    | |         |
[INFO] [stdout]    | |_________impl Future<Output = bareun_rs::tokenizer::SegResult>
[INFO] [stdout]    |           Vec<Vec<String>>
[INFO] [stdout]    |
[INFO] [stdout]    = note: this error originates in the macro `assert_eq` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0277]: `impl Future<Output = bareun_rs::tokenizer::SegResult>` doesn't implement `Debug`
[INFO] [stdout]   --> tests/test_tokenizer.rs:15:9
[INFO] [stdout]    |
[INFO] [stdout] 15 | /         assert_eq!(
[INFO] [stdout] 16 | |             tokenizer.await.seg(TEST_STR, false),
[INFO] [stdout] 17 | |             vec![
[INFO] [stdout] 18 | |                 vec!["오늘".to_string(), "은".to_string()],
[INFO] [stdout] ...  |
[INFO] [stdout] 29 | |         );
[INFO] [stdout]    | |_________^ the trait `Debug` is not implemented for `impl Future<Output = bareun_rs::tokenizer::SegResult>`
[INFO] [stdout]    |
[INFO] [stdout]    = note: this error originates in the macro `assert_eq` (in Nightly builds, run with -Z macro-backtrace for more info)
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> tests/test_tokenizer.rs:34:93
[INFO] [stdout]     |
[INFO] [stdout]  34 |         let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757);
[INFO] [stdout]     |                         -------------- arguments to this function are incorrect             ^^^^ expected `Option<i32>`, found integer
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected enum `Option<i32>`
[INFO] [stdout]                found type `{integer}`
[INFO] [stdout] note: associated function defined here
[INFO] [stdout]    --> /opt/rustwide/workdir/src/tokenizer.rs:322:18
[INFO] [stdout]     |
[INFO] [stdout] 322 |     pub async fn new(apikey: &str, host: &str, port: Option<i32>) -> Self {
[INFO] [stdout]     |                  ^^^
[INFO] [stdout] help: try wrapping the expression in `Some`
[INFO] [stdout]     |
[INFO] [stdout]  34 |         let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757));
[INFO] [stdout]     |                                                                                             +++++    +
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `seg` found for opaque type `impl Future<Output = Tokenizer>` in the current scope
[INFO] [stdout]   --> tests/test_tokenizer.rs:36:23
[INFO] [stdout]    |
[INFO] [stdout] 36 |             tokenizer.seg(TEST_STR, true, true),
[INFO] [stdout]    |                       ^^^ method not found in `impl Future<Output = Tokenizer>`
[INFO] [stdout]    |
[INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output`
[INFO] [stdout]    |
[INFO] [stdout] 36 |             tokenizer.await.seg(TEST_STR, true, true),
[INFO] [stdout]    |                       ++++++
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> tests/test_tokenizer.rs:54:93
[INFO] [stdout]     |
[INFO] [stdout]  54 |         let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757);
[INFO] [stdout]     |                         -------------- arguments to this function are incorrect             ^^^^ expected `Option<i32>`, found integer
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected enum `Option<i32>`
[INFO] [stdout]                found type `{integer}`
[INFO] [stdout] note: associated function defined here
[INFO] [stdout]    --> /opt/rustwide/workdir/src/tokenizer.rs:322:18
[INFO] [stdout]     |
[INFO] [stdout] 322 |     pub async fn new(apikey: &str, host: &str, port: Option<i32>) -> Self {
[INFO] [stdout]     |                  ^^^
[INFO] [stdout] help: try wrapping the expression in `Some`
[INFO] [stdout]     |
[INFO] [stdout]  54 |         let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757));
[INFO] [stdout]     |                                                                                             +++++    +
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `seg` found for opaque type `impl Future<Output = Tokenizer>` in the current scope
[INFO] [stdout]   --> tests/test_tokenizer.rs:55:30
[INFO] [stdout]    |
[INFO] [stdout] 55 |         let temp = tokenizer.seg(TEST_STR, true);
[INFO] [stdout]    |                              ^^^ method not found in `impl Future<Output = Tokenizer>`
[INFO] [stdout]    |
[INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output`
[INFO] [stdout]    |
[INFO] [stdout] 55 |         let temp = tokenizer.await.seg(TEST_STR, true);
[INFO] [stdout]    |                              ++++++
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> tests/test_tokenizer.rs:76:93
[INFO] [stdout]     |
[INFO] [stdout]  76 |         let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757);
[INFO] [stdout]     |                         -------------- arguments to this function are incorrect             ^^^^ expected `Option<i32>`, found integer
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected enum `Option<i32>`
[INFO] [stdout]                found type `{integer}`
[INFO] [stdout] note: associated function defined here
[INFO] [stdout]    --> /opt/rustwide/workdir/src/tokenizer.rs:322:18
[INFO] [stdout]     |
[INFO] [stdout] 322 |     pub async fn new(apikey: &str, host: &str, port: Option<i32>) -> Self {
[INFO] [stdout]     |                  ^^^
[INFO] [stdout] help: try wrapping the expression in `Some`
[INFO] [stdout]     |
[INFO] [stdout]  76 |         let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757));
[INFO] [stdout]     |                                                                                             +++++    +
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `seg` found for opaque type `impl Future<Output = Tokenizer>` in the current scope
[INFO] [stdout]   --> tests/test_tokenizer.rs:78:23
[INFO] [stdout]    |
[INFO] [stdout] 78 |             tokenizer.seg(TEST_STR),
[INFO] [stdout]    |                       ^^^ method not found in `impl Future<Output = Tokenizer>`
[INFO] [stdout]    |
[INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output`
[INFO] [stdout]    |
[INFO] [stdout] 78 |             tokenizer.await.seg(TEST_STR),
[INFO] [stdout]    |                       ++++++
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> tests/test_tokenizer.rs:96:93
[INFO] [stdout]     |
[INFO] [stdout]  96 |         let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757);
[INFO] [stdout]     |                         -------------- arguments to this function are incorrect             ^^^^ expected `Option<i32>`, found integer
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected enum `Option<i32>`
[INFO] [stdout]                found type `{integer}`
[INFO] [stdout] note: associated function defined here
[INFO] [stdout]    --> /opt/rustwide/workdir/src/tokenizer.rs:322:18
[INFO] [stdout]     |
[INFO] [stdout] 322 |     pub async fn new(apikey: &str, host: &str, port: Option<i32>) -> Self {
[INFO] [stdout]     |                  ^^^
[INFO] [stdout] help: try wrapping the expression in `Some`
[INFO] [stdout]     |
[INFO] [stdout]  96 |         let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757));
[INFO] [stdout]     |                                                                                             +++++    +
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `nouns` found for opaque type `impl Future<Output = Tokenizer>` in the current scope
[INFO] [stdout]   --> tests/test_tokenizer.rs:98:23
[INFO] [stdout]    |
[INFO] [stdout] 98 |             tokenizer.nouns(TEST_STR),
[INFO] [stdout]    |                       ^^^^^ method not found in `impl Future<Output = Tokenizer>`
[INFO] [stdout]    |
[INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output`
[INFO] [stdout]    |
[INFO] [stdout] 98 |             tokenizer.await.nouns(TEST_STR),
[INFO] [stdout]    |                       ++++++
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> tests/test_tokenizer.rs:105:93
[INFO] [stdout]     |
[INFO] [stdout] 105 |         let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757);
[INFO] [stdout]     |                         -------------- arguments to this function are incorrect             ^^^^ expected `Option<i32>`, found integer
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected enum `Option<i32>`
[INFO] [stdout]                found type `{integer}`
[INFO] [stdout] note: associated function defined here
[INFO] [stdout]    --> /opt/rustwide/workdir/src/tokenizer.rs:322:18
[INFO] [stdout]     |
[INFO] [stdout] 322 |     pub async fn new(apikey: &str, host: &str, port: Option<i32>) -> Self {
[INFO] [stdout]     |                  ^^^
[INFO] [stdout] help: try wrapping the expression in `Some`
[INFO] [stdout]     |
[INFO] [stdout] 105 |         let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757));
[INFO] [stdout]     |                                                                                             +++++    +
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `tokenize` found for opaque type `impl Future<Output = Tokenizer>` in the current scope
[INFO] [stdout]    --> tests/test_tokenizer.rs:106:35
[INFO] [stdout]     |
[INFO] [stdout] 106 |         let tokenized = tokenizer.tokenize(TEST_STR);
[INFO] [stdout]     |                                   ^^^^^^^^ method not found in `impl Future<Output = Tokenizer>`
[INFO] [stdout]     |
[INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output`
[INFO] [stdout]     |
[INFO] [stdout] 106 |         let tokenized = tokenizer.await.tokenize(TEST_STR);
[INFO] [stdout]     |                                   ++++++
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> tests/test_tokenizer.rs:115:93
[INFO] [stdout]     |
[INFO] [stdout] 115 |         let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757);
[INFO] [stdout]     |                         -------------- arguments to this function are incorrect             ^^^^ expected `Option<i32>`, found integer
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected enum `Option<i32>`
[INFO] [stdout]                found type `{integer}`
[INFO] [stdout] note: associated function defined here
[INFO] [stdout]    --> /opt/rustwide/workdir/src/tokenizer.rs:322:18
[INFO] [stdout]     |
[INFO] [stdout] 322 |     pub async fn new(apikey: &str, host: &str, port: Option<i32>) -> Self {
[INFO] [stdout]     |                  ^^^
[INFO] [stdout] help: try wrapping the expression in `Some`
[INFO] [stdout]     |
[INFO] [stdout] 115 |         let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757));
[INFO] [stdout]     |                                                                                             +++++    +
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `Error`
[INFO] [stdout]  --> src/lang_service_client.rs:3:42
[INFO] [stdout]   |
[INFO] [stdout] 3 | use tonic::transport::{channel, Channel, Error};
[INFO] [stdout]   |                                          ^^^^^
[INFO] [stdout]   |
[INFO] [stdout]   = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `Tag`
[INFO] [stdout]  --> src/tagger.rs:3:43
[INFO] [stdout]   |
[INFO] [stdout] 3 | use crate::bareun::morpheme::{OutOfVocab, Tag};
[INFO] [stdout]   |                                           ^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `MAX_MESSAGE_LENGTH`
[INFO] [stdout]  --> src/tagger.rs:7:63
[INFO] [stdout]   |
[INFO] [stdout] 7 | use crate::lang_service_client::{BareunLanguageServiceClient, MAX_MESSAGE_LENGTH};
[INFO] [stdout]   |                                                               ^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `tonic::transport::Endpoint`
[INFO] [stdout]   --> src/tagger.rs:11:5
[INFO] [stdout]    |
[INFO] [stdout] 11 | use tonic::transport::Endpoint;
[INFO] [stdout]    |     ^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `std::collections::HashMap`
[INFO] [stdout]  --> src/tokenizer.rs:1:5
[INFO] [stdout]   |
[INFO] [stdout] 1 | use std::collections::HashMap;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused imports: `SegmentToken` and `TextSpan`
[INFO] [stdout]  --> src/tokenizer.rs:3:47
[INFO] [stdout]   |
[INFO] [stdout] 3 | use crate::bareun::{Segment, SegmentSentence, SegmentToken, TextSpan, TokenizeResponse};
[INFO] [stdout]   |                                               ^^^^^^^^^^^^  ^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `MAX_MESSAGE_LENGTH`
[INFO] [stdout]  --> src/tokenizer.rs:4:63
[INFO] [stdout]   |
[INFO] [stdout] 4 | use crate::lang_service_client::{BareunLanguageServiceClient, MAX_MESSAGE_LENGTH};
[INFO] [stdout]   |                                                               ^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused imports: `Deserialize` and `Serialize`
[INFO] [stdout]  --> src/tokenizer.rs:5:13
[INFO] [stdout]   |
[INFO] [stdout] 5 | use serde::{Deserialize, Serialize};
[INFO] [stdout]   |             ^^^^^^^^^^^  ^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `serde_json::json`
[INFO] [stdout]  --> src/tokenizer.rs:7:5
[INFO] [stdout]   |
[INFO] [stdout] 7 | use serde_json::json;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: unused import: `tonic::transport::Endpoint`
[INFO] [stdout]  --> src/tokenizer.rs:8:5
[INFO] [stdout]   |
[INFO] [stdout] 8 | use tonic::transport::Endpoint;
[INFO] [stdout]   |     ^^^^^^^^^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `tokenize` found for opaque type `impl Future<Output = Tokenizer>` in the current scope
[INFO] [stdout]    --> tests/test_tokenizer.rs:116:35
[INFO] [stdout]     |
[INFO] [stdout] 116 |         let tokenized = tokenizer.tokenize(TEST_STR);
[INFO] [stdout]     |                                   ^^^^^^^^ method not found in `impl Future<Output = Tokenizer>`
[INFO] [stdout]     |
[INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output`
[INFO] [stdout]     |
[INFO] [stdout] 116 |         let tokenized = tokenizer.await.tokenize(TEST_STR);
[INFO] [stdout]     |                                   ++++++
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> tests/test_tokenizer.rs:125:93
[INFO] [stdout]     |
[INFO] [stdout] 125 |         let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757);
[INFO] [stdout]     |                         -------------- arguments to this function are incorrect             ^^^^ expected `Option<i32>`, found integer
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected enum `Option<i32>`
[INFO] [stdout]                found type `{integer}`
[INFO] [stdout] note: associated function defined here
[INFO] [stdout]    --> /opt/rustwide/workdir/src/tokenizer.rs:322:18
[INFO] [stdout]     |
[INFO] [stdout] 322 |     pub async fn new(apikey: &str, host: &str, port: Option<i32>) -> Self {
[INFO] [stdout]     |                  ^^^
[INFO] [stdout] help: try wrapping the expression in `Some`
[INFO] [stdout]     |
[INFO] [stdout] 125 |         let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757));
[INFO] [stdout]     |                                                                                             +++++    +
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `tokenize` found for opaque type `impl Future<Output = Tokenizer>` in the current scope
[INFO] [stdout]    --> tests/test_tokenizer.rs:126:35
[INFO] [stdout]     |
[INFO] [stdout] 126 |         let tokenized = tokenizer.tokenize(TEST_STR);
[INFO] [stdout]     |                                   ^^^^^^^^ method not found in `impl Future<Output = Tokenizer>`
[INFO] [stdout]     |
[INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output`
[INFO] [stdout]     |
[INFO] [stdout] 126 |         let tokenized = tokenizer.await.tokenize(TEST_STR);
[INFO] [stdout]     |                                   ++++++
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> tests/test_tokenizer.rs:132:93
[INFO] [stdout]     |
[INFO] [stdout] 132 |         let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757);
[INFO] [stdout]     |                         -------------- arguments to this function are incorrect             ^^^^ expected `Option<i32>`, found integer
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected enum `Option<i32>`
[INFO] [stdout]                found type `{integer}`
[INFO] [stdout] note: associated function defined here
[INFO] [stdout]    --> /opt/rustwide/workdir/src/tokenizer.rs:322:18
[INFO] [stdout]     |
[INFO] [stdout] 322 |     pub async fn new(apikey: &str, host: &str, port: Option<i32>) -> Self {
[INFO] [stdout]     |                  ^^^
[INFO] [stdout] help: try wrapping the expression in `Some`
[INFO] [stdout]     |
[INFO] [stdout] 132 |         let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757));
[INFO] [stdout]     |                                                                                             +++++    +
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `tokenize` found for opaque type `impl Future<Output = Tokenizer>` in the current scope
[INFO] [stdout]    --> tests/test_tokenizer.rs:133:35
[INFO] [stdout]     |
[INFO] [stdout] 133 |         let tokenized = tokenizer.tokenize(TEST_STR);
[INFO] [stdout]     |                                   ^^^^^^^^ method not found in `impl Future<Output = Tokenizer>`
[INFO] [stdout]     |
[INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output`
[INFO] [stdout]     |
[INFO] [stdout] 133 |         let tokenized = tokenizer.await.tokenize(TEST_STR);
[INFO] [stdout]     |                                   ++++++
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> tests/test_tokenizer.rs:139:93
[INFO] [stdout]     |
[INFO] [stdout] 139 |         let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757);
[INFO] [stdout]     |                         -------------- arguments to this function are incorrect             ^^^^ expected `Option<i32>`, found integer
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected enum `Option<i32>`
[INFO] [stdout]                found type `{integer}`
[INFO] [stdout] note: associated function defined here
[INFO] [stdout]    --> /opt/rustwide/workdir/src/tokenizer.rs:322:18
[INFO] [stdout]     |
[INFO] [stdout] 322 |     pub async fn new(apikey: &str, host: &str, port: Option<i32>) -> Self {
[INFO] [stdout]     |                  ^^^
[INFO] [stdout] help: try wrapping the expression in `Some`
[INFO] [stdout]     |
[INFO] [stdout] 139 |         let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757));
[INFO] [stdout]     |                                                                                             +++++    +
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `tokenize` found for opaque type `impl Future<Output = Tokenizer>` in the current scope
[INFO] [stdout]    --> tests/test_tokenizer.rs:140:35
[INFO] [stdout]     |
[INFO] [stdout] 140 |         let tokenized = tokenizer.tokenize(TEST_STR);
[INFO] [stdout]     |                                   ^^^^^^^^ method not found in `impl Future<Output = Tokenizer>`
[INFO] [stdout]     |
[INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output`
[INFO] [stdout]     |
[INFO] [stdout] 140 |         let tokenized = tokenizer.await.tokenize(TEST_STR);
[INFO] [stdout]     |                                   ++++++
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0308]: mismatched types
[INFO] [stdout]    --> tests/test_tokenizer.rs:149:93
[INFO] [stdout]     |
[INFO] [stdout] 149 |         let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757);
[INFO] [stdout]     |                         -------------- arguments to this function are incorrect             ^^^^ expected `Option<i32>`, found integer
[INFO] [stdout]     |
[INFO] [stdout]     = note: expected enum `Option<i32>`
[INFO] [stdout]                found type `{integer}`
[INFO] [stdout] note: associated function defined here
[INFO] [stdout]    --> /opt/rustwide/workdir/src/tokenizer.rs:322:18
[INFO] [stdout]     |
[INFO] [stdout] 322 |     pub async fn new(apikey: &str, host: &str, port: Option<i32>) -> Self {
[INFO] [stdout]     |                  ^^^
[INFO] [stdout] help: try wrapping the expression in `Some`
[INFO] [stdout]     |
[INFO] [stdout] 149 |         let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757));
[INFO] [stdout]     |                                                                                             +++++    +
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] error[E0599]: no method named `tokenize` found for opaque type `impl Future<Output = Tokenizer>` in the current scope
[INFO] [stdout]    --> tests/test_tokenizer.rs:150:35
[INFO] [stdout]     |
[INFO] [stdout] 150 |         let tokenized = tokenizer.tokenize(TEST_STR);
[INFO] [stdout]     |                                   ^^^^^^^^ method not found in `impl Future<Output = Tokenizer>`
[INFO] [stdout]     |
[INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output`
[INFO] [stdout]     |
[INFO] [stdout] 150 |         let tokenized = tokenizer.await.tokenize(TEST_STR);
[INFO] [stdout]     |                                   ++++++
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] Some errors have detailed explanations: E0061, E0277, E0308, E0369, E0599.
[INFO] [stdout] 
[INFO] [stdout] For more information about an error, try `rustc --explain E0061`.
[INFO] [stdout] 
[INFO] [stderr] error: could not compile `bareun_rs` (test "test_tokenizer") due to 23 previous errors; 1 warning emitted
[INFO] [stderr] warning: build failed, waiting for other jobs to finish...
[INFO] [stdout] warning: function `pb_map_to_set` is never used
[INFO] [stdout]   --> src/custom_dict.rs:39:8
[INFO] [stdout]    |
[INFO] [stdout] 39 | pub fn pb_map_to_set(ds: &DictSet) -> HashSet<String> {
[INFO] [stdout]    |        ^^^^^^^^^^^^^
[INFO] [stdout]    |
[INFO] [stdout]    = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: constant `MAX_MESSAGE_LENGTH` is never used
[INFO] [stdout]   --> src/lang_service_client.rs:12:11
[INFO] [stdout]    |
[INFO] [stdout] 12 | pub const MAX_MESSAGE_LENGTH: usize = 100 * 1024 * 1024;
[INFO] [stdout]    |           ^^^^^^^^^^^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: field `metadata` is never read
[INFO] [stdout]   --> src/lang_service_client.rs:19:5
[INFO] [stdout]    |
[INFO] [stdout] 13 | pub struct BareunLanguageServiceClient {
[INFO] [stdout]    |            --------------------------- field in this struct
[INFO] [stdout] ...
[INFO] [stdout] 19 |     metadata: HashMap<String, String>,
[INFO] [stdout]    |     ^^^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: field `phrase` is never read
[INFO] [stdout]   --> src/tagger.rs:17:5
[INFO] [stdout]    |
[INFO] [stdout] 12 | pub struct Tagged {
[INFO] [stdout]    |            ------ field in this struct
[INFO] [stdout] ...
[INFO] [stdout] 17 |     phrase: String,
[INFO] [stdout]    |     ^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] [stdout] warning: field `phrase` is never read
[INFO] [stdout]   --> src/tokenizer.rs:18:5
[INFO] [stdout]    |
[INFO] [stdout] 13 | pub struct Tokenized {
[INFO] [stdout]    |            --------- field in this struct
[INFO] [stdout] ...
[INFO] [stdout] 18 |     phrase: String,
[INFO] [stdout]    |     ^^^^^^
[INFO] [stdout] 
[INFO] [stdout] 
[INFO] running `Command { std: "docker" "inspect" "0eeefc8044c2d76ffa371c7b39f98412d9e9975d99799c6020f95b088566d518", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "0eeefc8044c2d76ffa371c7b39f98412d9e9975d99799c6020f95b088566d518", kill_on_drop: false }`
[INFO] [stdout] 0eeefc8044c2d76ffa371c7b39f98412d9e9975d99799c6020f95b088566d518
