[INFO] fetching crate bareun_rs 0.1.0... [INFO] testing bareun_rs-0.1.0 against master#cdb45c87e2cd43495379f7e867e3cc15dcee9f93 for pr-145838-1 [INFO] extracting crate bareun_rs 0.1.0 into /workspace/builds/worker-1-tc1/source [INFO] started tweaking crates.io crate bareun_rs 0.1.0 [INFO] finished tweaking crates.io crate bareun_rs 0.1.0 [INFO] tweaked toml for crates.io crate bareun_rs 0.1.0 written to /workspace/builds/worker-1-tc1/source/Cargo.toml [INFO] validating manifest of crates.io crate bareun_rs 0.1.0 on toolchain cdb45c87e2cd43495379f7e867e3cc15dcee9f93 [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+cdb45c87e2cd43495379f7e867e3cc15dcee9f93" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }` [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+cdb45c87e2cd43495379f7e867e3cc15dcee9f93" "generate-lockfile" "--manifest-path" "Cargo.toml", kill_on_drop: false }` [INFO] [stderr] Updating crates.io index [INFO] [stderr] Locking 140 packages to latest compatible versions [INFO] [stderr] Adding prost v0.12.6 (available: v0.14.1) [INFO] [stderr] Adding tonic v0.11.0 (available: v0.14.1) [INFO] [stderr] Adding tonic-build v0.11.0 (available: v0.14.1) [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+cdb45c87e2cd43495379f7e867e3cc15dcee9f93" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }` [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-1-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-1-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:7ad1b28ee6f5f7f699f6cf7015098d6ccdd96d6f2d78dd06228f5b4c9faf309c" "/opt/rustwide/cargo-home/bin/cargo" "+cdb45c87e2cd43495379f7e867e3cc15dcee9f93" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }` [INFO] [stdout] d4fc3218f85173a94d69976cac816d61ba906da116d74a33f0ed069a582218b6 [INFO] running `Command { std: "docker" "start" "-a" "d4fc3218f85173a94d69976cac816d61ba906da116d74a33f0ed069a582218b6", kill_on_drop: false }` [INFO] running `Command { std: "docker" "inspect" "d4fc3218f85173a94d69976cac816d61ba906da116d74a33f0ed069a582218b6", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "d4fc3218f85173a94d69976cac816d61ba906da116d74a33f0ed069a582218b6", kill_on_drop: false }` [INFO] [stdout] d4fc3218f85173a94d69976cac816d61ba906da116d74a33f0ed069a582218b6 [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-1-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-1-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:7ad1b28ee6f5f7f699f6cf7015098d6ccdd96d6f2d78dd06228f5b4c9faf309c" "/opt/rustwide/cargo-home/bin/cargo" "+cdb45c87e2cd43495379f7e867e3cc15dcee9f93" "build" "--frozen" "--message-format=json", kill_on_drop: false }` [INFO] [stdout] b4eb94924f00702697c774fb90edb242b9eb065dbaf152cafa66bbc7fb5ef987 [INFO] running `Command { std: "docker" "start" "-a" "b4eb94924f00702697c774fb90edb242b9eb065dbaf152cafa66bbc7fb5ef987", kill_on_drop: false }` [INFO] [stderr] Compiling parking_lot_core v0.9.11 [INFO] [stderr] Compiling smallvec v1.15.1 [INFO] [stderr] Compiling hashbrown v0.15.5 [INFO] [stderr] Compiling equivalent v1.0.2 [INFO] [stderr] Compiling either v1.15.0 [INFO] [stderr] Compiling lock_api v0.4.13 [INFO] [stderr] Compiling anyhow v1.0.99 [INFO] [stderr] Compiling rustversion v1.0.22 [INFO] [stderr] Compiling http v0.2.12 [INFO] [stderr] Compiling futures-util v0.3.31 [INFO] [stderr] Compiling libc v0.2.175 [INFO] [stderr] Compiling syn v2.0.106 [INFO] [stderr] Compiling signal-hook-registry v1.4.6 [INFO] [stderr] Compiling rustix v1.0.8 [INFO] [stderr] Compiling getrandom v0.3.3 [INFO] [stderr] Compiling bytes v1.10.1 [INFO] [stderr] Compiling bitflags v2.9.3 [INFO] [stderr] Compiling prettyplease v0.2.37 [INFO] [stderr] Compiling itertools v0.12.1 [INFO] [stderr] Compiling linux-raw-sys v0.9.4 [INFO] [stderr] Compiling regex-syntax v0.8.6 [INFO] [stderr] Compiling indexmap v1.9.3 [INFO] [stderr] Compiling fastrand v2.3.0 [INFO] [stderr] Compiling serde v1.0.219 [INFO] [stderr] Compiling parking_lot v0.12.4 [INFO] [stderr] Compiling fixedbitset v0.4.2 [INFO] [stderr] Compiling indexmap v2.11.0 [INFO] [stderr] Compiling http-body v0.4.6 [INFO] [stderr] Compiling socket2 v0.5.10 [INFO] [stderr] Compiling futures-channel v0.3.31 [INFO] [stderr] Compiling log v0.4.27 [INFO] [stderr] Compiling hashbrown v0.12.3 [INFO] [stderr] Compiling heck v0.5.0 [INFO] [stderr] Compiling multimap v0.10.1 [INFO] [stderr] Compiling httpdate v1.0.3 [INFO] [stderr] Compiling rand v0.8.5 [INFO] [stderr] Compiling mime v0.3.17 [INFO] [stderr] Compiling sync_wrapper v0.1.2 [INFO] [stderr] Compiling matchit v0.7.3 [INFO] [stderr] Compiling base64 v0.21.7 [INFO] [stderr] Compiling axum-core v0.3.4 [INFO] [stderr] Compiling axum v0.6.20 [INFO] [stderr] Compiling regex-automata v0.4.10 [INFO] [stderr] Compiling petgraph v0.6.5 [INFO] [stderr] Compiling tempfile v3.21.0 [INFO] [stderr] Compiling regex v1.11.2 [INFO] [stderr] Compiling tokio-macros v2.5.0 [INFO] [stderr] Compiling prost-derive v0.12.6 [INFO] [stderr] Compiling tracing-attributes v0.1.30 [INFO] [stderr] Compiling serde_derive v1.0.219 [INFO] [stderr] Compiling pin-project-internal v1.1.10 [INFO] [stderr] Compiling async-trait v0.1.89 [INFO] [stderr] Compiling async-stream-impl v0.3.6 [INFO] [stderr] Compiling async-stream v0.3.6 [INFO] [stderr] Compiling tokio v1.47.1 [INFO] [stderr] Compiling pin-project v1.1.10 [INFO] [stderr] Compiling tracing v0.1.41 [INFO] [stderr] Compiling prost v0.12.6 [INFO] [stderr] Compiling prost-types v0.12.6 [INFO] [stderr] Compiling prost-build v0.12.6 [INFO] [stderr] Compiling tonic-build v0.11.0 [INFO] [stderr] Compiling bareun_rs v0.1.0 (/opt/rustwide/workdir) [INFO] [stderr] Compiling tokio-util v0.7.16 [INFO] [stderr] Compiling tokio-io-timeout v1.2.1 [INFO] [stderr] Compiling tokio-stream v0.1.17 [INFO] [stderr] Compiling h2 v0.3.27 [INFO] [stderr] Compiling tower v0.4.13 [INFO] [stderr] Compiling serde_json v1.0.143 [INFO] [stderr] Compiling hyper v0.14.32 [INFO] [stderr] Compiling hyper-timeout v0.4.1 [INFO] [stderr] Compiling tonic v0.11.0 [INFO] [stdout] warning: unused import: `Error` [INFO] [stdout] --> src/lang_service_client.rs:3:42 [INFO] [stdout] | [INFO] [stdout] 3 | use tonic::transport::{channel, Channel, Error}; [INFO] [stdout] | ^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `Tag` [INFO] [stdout] --> src/tagger.rs:3:43 [INFO] [stdout] | [INFO] [stdout] 3 | use crate::bareun::morpheme::{OutOfVocab, Tag}; [INFO] [stdout] | ^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `MAX_MESSAGE_LENGTH` [INFO] [stdout] --> src/tagger.rs:7:63 [INFO] [stdout] | [INFO] [stdout] 7 | use crate::lang_service_client::{BareunLanguageServiceClient, MAX_MESSAGE_LENGTH}; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `tonic::transport::Endpoint` [INFO] [stdout] --> src/tagger.rs:11:5 [INFO] [stdout] | [INFO] [stdout] 11 | use tonic::transport::Endpoint; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `std::collections::HashMap` [INFO] [stdout] --> src/tokenizer.rs:1:5 [INFO] [stdout] | [INFO] [stdout] 1 | use std::collections::HashMap; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused imports: `SegmentToken` and `TextSpan` [INFO] [stdout] --> src/tokenizer.rs:3:47 [INFO] [stdout] | [INFO] [stdout] 3 | use crate::bareun::{Segment, SegmentSentence, SegmentToken, TextSpan, TokenizeResponse}; [INFO] [stdout] | ^^^^^^^^^^^^ ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `MAX_MESSAGE_LENGTH` [INFO] [stdout] --> src/tokenizer.rs:4:63 [INFO] [stdout] | [INFO] [stdout] 4 | use crate::lang_service_client::{BareunLanguageServiceClient, MAX_MESSAGE_LENGTH}; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused imports: `Deserialize` and `Serialize` [INFO] [stdout] --> src/tokenizer.rs:5:13 [INFO] [stdout] | [INFO] [stdout] 5 | use serde::{Deserialize, Serialize}; [INFO] [stdout] | ^^^^^^^^^^^ ^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `serde_json::json` [INFO] [stdout] --> src/tokenizer.rs:7:5 [INFO] [stdout] | [INFO] [stdout] 7 | use serde_json::json; [INFO] [stdout] | ^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `tonic::transport::Endpoint` [INFO] [stdout] --> src/tokenizer.rs:8:5 [INFO] [stdout] | [INFO] [stdout] 8 | use tonic::transport::Endpoint; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `pb_map_to_set` is never used [INFO] [stdout] --> src/custom_dict.rs:39:8 [INFO] [stdout] | [INFO] [stdout] 39 | pub fn pb_map_to_set(ds: &DictSet) -> HashSet { [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: constant `MAX_MESSAGE_LENGTH` is never used [INFO] [stdout] --> src/lang_service_client.rs:12:11 [INFO] [stdout] | [INFO] [stdout] 12 | pub const MAX_MESSAGE_LENGTH: usize = 100 * 1024 * 1024; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `metadata` is never read [INFO] [stdout] --> src/lang_service_client.rs:19:5 [INFO] [stdout] | [INFO] [stdout] 13 | pub struct BareunLanguageServiceClient { [INFO] [stdout] | --------------------------- field in this struct [INFO] [stdout] ... [INFO] [stdout] 19 | metadata: HashMap, [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `phrase` is never read [INFO] [stdout] --> src/tagger.rs:17:5 [INFO] [stdout] | [INFO] [stdout] 12 | pub struct Tagged { [INFO] [stdout] | ------ field in this struct [INFO] [stdout] ... [INFO] [stdout] 17 | phrase: String, [INFO] [stdout] | ^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `phrase` is never read [INFO] [stdout] --> src/tokenizer.rs:18:5 [INFO] [stdout] | [INFO] [stdout] 13 | pub struct Tokenized { [INFO] [stdout] | --------- field in this struct [INFO] [stdout] ... [INFO] [stdout] 18 | phrase: String, [INFO] [stdout] | ^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] Finished `dev` profile [unoptimized + debuginfo] target(s) in 44.34s [INFO] running `Command { std: "docker" "inspect" "b4eb94924f00702697c774fb90edb242b9eb065dbaf152cafa66bbc7fb5ef987", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "b4eb94924f00702697c774fb90edb242b9eb065dbaf152cafa66bbc7fb5ef987", kill_on_drop: false }` [INFO] [stdout] b4eb94924f00702697c774fb90edb242b9eb065dbaf152cafa66bbc7fb5ef987 [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-1-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-1-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:7ad1b28ee6f5f7f699f6cf7015098d6ccdd96d6f2d78dd06228f5b4c9faf309c" "/opt/rustwide/cargo-home/bin/cargo" "+cdb45c87e2cd43495379f7e867e3cc15dcee9f93" "test" "--frozen" "--no-run" "--message-format=json", kill_on_drop: false }` [INFO] [stdout] 08240f2ce1cfac83e865ed3b68bf6a2239852106aabf2b50736decfcdfe4b76d [INFO] running `Command { std: "docker" "start" "-a" "08240f2ce1cfac83e865ed3b68bf6a2239852106aabf2b50736decfcdfe4b76d", kill_on_drop: false }` [INFO] [stdout] warning: unused import: `Error` [INFO] [stdout] --> src/lang_service_client.rs:3:42 [INFO] [stdout] | [INFO] [stdout] 3 | use tonic::transport::{channel, Channel, Error}; [INFO] [stdout] | ^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `Tag` [INFO] [stdout] --> src/tagger.rs:3:43 [INFO] [stdout] | [INFO] [stdout] 3 | use crate::bareun::morpheme::{OutOfVocab, Tag}; [INFO] [stdout] | ^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `MAX_MESSAGE_LENGTH` [INFO] [stdout] --> src/tagger.rs:7:63 [INFO] [stdout] | [INFO] [stdout] 7 | use crate::lang_service_client::{BareunLanguageServiceClient, MAX_MESSAGE_LENGTH}; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `tonic::transport::Endpoint` [INFO] [stdout] --> src/tagger.rs:11:5 [INFO] [stdout] | [INFO] [stdout] 11 | use tonic::transport::Endpoint; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `std::collections::HashMap` [INFO] [stdout] --> src/tokenizer.rs:1:5 [INFO] [stdout] | [INFO] [stdout] 1 | use std::collections::HashMap; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused imports: `SegmentToken` and `TextSpan` [INFO] [stdout] --> src/tokenizer.rs:3:47 [INFO] [stdout] | [INFO] [stdout] 3 | use crate::bareun::{Segment, SegmentSentence, SegmentToken, TextSpan, TokenizeResponse}; [INFO] [stdout] | ^^^^^^^^^^^^ ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `MAX_MESSAGE_LENGTH` [INFO] [stdout] --> src/tokenizer.rs:4:63 [INFO] [stdout] | [INFO] [stdout] 4 | use crate::lang_service_client::{BareunLanguageServiceClient, MAX_MESSAGE_LENGTH}; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused imports: `Deserialize` and `Serialize` [INFO] [stdout] --> src/tokenizer.rs:5:13 [INFO] [stdout] | [INFO] [stdout] 5 | use serde::{Deserialize, Serialize}; [INFO] [stdout] | ^^^^^^^^^^^ ^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `serde_json::json` [INFO] [stdout] --> src/tokenizer.rs:7:5 [INFO] [stdout] | [INFO] [stdout] 7 | use serde_json::json; [INFO] [stdout] | ^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `tonic::transport::Endpoint` [INFO] [stdout] --> src/tokenizer.rs:8:5 [INFO] [stdout] | [INFO] [stdout] 8 | use tonic::transport::Endpoint; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `pb_map_to_set` is never used [INFO] [stdout] --> src/custom_dict.rs:39:8 [INFO] [stdout] | [INFO] [stdout] 39 | pub fn pb_map_to_set(ds: &DictSet) -> HashSet { [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: constant `MAX_MESSAGE_LENGTH` is never used [INFO] [stdout] --> src/lang_service_client.rs:12:11 [INFO] [stdout] | [INFO] [stdout] 12 | pub const MAX_MESSAGE_LENGTH: usize = 100 * 1024 * 1024; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `metadata` is never read [INFO] [stdout] --> src/lang_service_client.rs:19:5 [INFO] [stdout] | [INFO] [stdout] 13 | pub struct BareunLanguageServiceClient { [INFO] [stdout] | --------------------------- field in this struct [INFO] [stdout] ... [INFO] [stdout] 19 | metadata: HashMap, [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `phrase` is never read [INFO] [stdout] --> src/tagger.rs:17:5 [INFO] [stdout] | [INFO] [stdout] 12 | pub struct Tagged { [INFO] [stdout] | ------ field in this struct [INFO] [stdout] ... [INFO] [stdout] 17 | phrase: String, [INFO] [stdout] | ^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `phrase` is never read [INFO] [stdout] --> src/tokenizer.rs:18:5 [INFO] [stdout] | [INFO] [stdout] 13 | pub struct Tokenized { [INFO] [stdout] | --------- field in this struct [INFO] [stdout] ... [INFO] [stdout] 18 | phrase: String, [INFO] [stdout] | ^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] Compiling bareun_rs v0.1.0 (/opt/rustwide/workdir) [INFO] [stdout] warning: unused import: `bareun_rs::Tokenized` [INFO] [stdout] --> tests/test_tokenizer.rs:3:9 [INFO] [stdout] | [INFO] [stdout] 3 | use bareun_rs::Tokenized; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `Error` [INFO] [stdout] --> src/lang_service_client.rs:3:42 [INFO] [stdout] | [INFO] [stdout] 3 | use tonic::transport::{channel, Channel, Error}; [INFO] [stdout] | ^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `Tag` [INFO] [stdout] --> src/tagger.rs:3:43 [INFO] [stdout] | [INFO] [stdout] 3 | use crate::bareun::morpheme::{OutOfVocab, Tag}; [INFO] [stdout] | ^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `MAX_MESSAGE_LENGTH` [INFO] [stdout] --> src/tagger.rs:7:63 [INFO] [stdout] | [INFO] [stdout] 7 | use crate::lang_service_client::{BareunLanguageServiceClient, MAX_MESSAGE_LENGTH}; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `tonic::transport::Endpoint` [INFO] [stdout] --> src/tagger.rs:11:5 [INFO] [stdout] | [INFO] [stdout] 11 | use tonic::transport::Endpoint; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `std::collections::HashMap` [INFO] [stdout] --> src/tokenizer.rs:1:5 [INFO] [stdout] | [INFO] [stdout] 1 | use std::collections::HashMap; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused imports: `SegmentToken` and `TextSpan` [INFO] [stdout] --> src/tokenizer.rs:3:47 [INFO] [stdout] | [INFO] [stdout] 3 | use crate::bareun::{Segment, SegmentSentence, SegmentToken, TextSpan, TokenizeResponse}; [INFO] [stdout] | ^^^^^^^^^^^^ ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `MAX_MESSAGE_LENGTH` [INFO] [stdout] --> src/tokenizer.rs:4:63 [INFO] [stdout] | [INFO] [stdout] 4 | use crate::lang_service_client::{BareunLanguageServiceClient, MAX_MESSAGE_LENGTH}; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused imports: `Deserialize` and `Serialize` [INFO] [stdout] --> src/tokenizer.rs:5:13 [INFO] [stdout] | [INFO] [stdout] 5 | use serde::{Deserialize, Serialize}; [INFO] [stdout] | ^^^^^^^^^^^ ^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `serde_json::json` [INFO] [stdout] --> src/tokenizer.rs:7:5 [INFO] [stdout] | [INFO] [stdout] 7 | use serde_json::json; [INFO] [stdout] | ^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `tonic::transport::Endpoint` [INFO] [stdout] --> src/tokenizer.rs:8:5 [INFO] [stdout] | [INFO] [stdout] 8 | use tonic::transport::Endpoint; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this method takes 4 arguments but 2 arguments were supplied [INFO] [stdout] --> tests/test_tokenizer.rs:16:29 [INFO] [stdout] | [INFO] [stdout] 16 | tokenizer.await.seg(TEST_STR, false), [INFO] [stdout] | ^^^----------------- two arguments of type `bool` and `bool` are missing [INFO] [stdout] | [INFO] [stdout] note: method defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:381:18 [INFO] [stdout] | [INFO] [stdout] 381 | pub async fn seg( [INFO] [stdout] | ^^^ [INFO] [stdout] help: provide the arguments [INFO] [stdout] | [INFO] [stdout] 16 | tokenizer.await.seg(TEST_STR, false, /* bool */, /* bool */), [INFO] [stdout] | ++++++++++++++++++++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0369]: binary operation `==` cannot be applied to type `impl Future` [INFO] [stdout] --> tests/test_tokenizer.rs:15:9 [INFO] [stdout] | [INFO] [stdout] 15 | / assert_eq!( [INFO] [stdout] 16 | | tokenizer.await.seg(TEST_STR, false), [INFO] [stdout] 17 | | vec![ [INFO] [stdout] 18 | | vec!["오늘".to_string(), "은".to_string()], [INFO] [stdout] ... | [INFO] [stdout] 29 | | ); [INFO] [stdout] | | ^ [INFO] [stdout] | | | [INFO] [stdout] | |_________impl Future [INFO] [stdout] | Vec> [INFO] [stdout] | [INFO] [stdout] = note: this error originates in the macro `assert_eq` (in Nightly builds, run with -Z macro-backtrace for more info) [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0277]: `impl Future` doesn't implement `Debug` [INFO] [stdout] --> tests/test_tokenizer.rs:15:9 [INFO] [stdout] | [INFO] [stdout] 15 | / assert_eq!( [INFO] [stdout] 16 | | tokenizer.await.seg(TEST_STR, false), [INFO] [stdout] 17 | | vec![ [INFO] [stdout] 18 | | vec!["오늘".to_string(), "은".to_string()], [INFO] [stdout] ... | [INFO] [stdout] 29 | | ); [INFO] [stdout] | |_________^ the trait `Debug` is not implemented for `impl Future` [INFO] [stdout] | [INFO] [stdout] = note: this error originates in the macro `assert_eq` (in Nightly builds, run with -Z macro-backtrace for more info) [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:34:93 [INFO] [stdout] | [INFO] [stdout] 34 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 34 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `seg` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:36:23 [INFO] [stdout] | [INFO] [stdout] 36 | tokenizer.seg(TEST_STR, true, true), [INFO] [stdout] | ^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 36 | tokenizer.await.seg(TEST_STR, true, true), [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:54:93 [INFO] [stdout] | [INFO] [stdout] 54 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 54 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `seg` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:55:30 [INFO] [stdout] | [INFO] [stdout] 55 | let temp = tokenizer.seg(TEST_STR, true); [INFO] [stdout] | ^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 55 | let temp = tokenizer.await.seg(TEST_STR, true); [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:76:93 [INFO] [stdout] | [INFO] [stdout] 76 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 76 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `seg` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:78:23 [INFO] [stdout] | [INFO] [stdout] 78 | tokenizer.seg(TEST_STR), [INFO] [stdout] | ^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 78 | tokenizer.await.seg(TEST_STR), [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:96:93 [INFO] [stdout] | [INFO] [stdout] 96 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 96 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `nouns` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:98:23 [INFO] [stdout] | [INFO] [stdout] 98 | tokenizer.nouns(TEST_STR), [INFO] [stdout] | ^^^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 98 | tokenizer.await.nouns(TEST_STR), [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:105:93 [INFO] [stdout] | [INFO] [stdout] 105 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 105 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `tokenize` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:106:35 [INFO] [stdout] | [INFO] [stdout] 106 | let tokenized = tokenizer.tokenize(TEST_STR); [INFO] [stdout] | ^^^^^^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 106 | let tokenized = tokenizer.await.tokenize(TEST_STR); [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:115:93 [INFO] [stdout] | [INFO] [stdout] 115 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 115 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `tokenize` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:116:35 [INFO] [stdout] | [INFO] [stdout] 116 | let tokenized = tokenizer.tokenize(TEST_STR); [INFO] [stdout] | ^^^^^^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 116 | let tokenized = tokenizer.await.tokenize(TEST_STR); [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:125:93 [INFO] [stdout] | [INFO] [stdout] 125 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 125 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `tokenize` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:126:35 [INFO] [stdout] | [INFO] [stdout] 126 | let tokenized = tokenizer.tokenize(TEST_STR); [INFO] [stdout] | ^^^^^^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 126 | let tokenized = tokenizer.await.tokenize(TEST_STR); [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:132:93 [INFO] [stdout] | [INFO] [stdout] 132 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 132 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `tokenize` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:133:35 [INFO] [stdout] | [INFO] [stdout] 133 | let tokenized = tokenizer.tokenize(TEST_STR); [INFO] [stdout] | ^^^^^^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 133 | let tokenized = tokenizer.await.tokenize(TEST_STR); [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:139:93 [INFO] [stdout] | [INFO] [stdout] 139 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 139 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `tokenize` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:140:35 [INFO] [stdout] | [INFO] [stdout] 140 | let tokenized = tokenizer.tokenize(TEST_STR); [INFO] [stdout] | ^^^^^^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 140 | let tokenized = tokenizer.await.tokenize(TEST_STR); [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:149:93 [INFO] [stdout] | [INFO] [stdout] 149 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 149 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `tokenize` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:150:35 [INFO] [stdout] | [INFO] [stdout] 150 | let tokenized = tokenizer.tokenize(TEST_STR); [INFO] [stdout] | ^^^^^^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 150 | let tokenized = tokenizer.await.tokenize(TEST_STR); [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] Some errors have detailed explanations: E0061, E0277, E0308, E0369, E0599. [INFO] [stdout] [INFO] [stdout] For more information about an error, try `rustc --explain E0061`. [INFO] [stdout] [INFO] [stderr] error: could not compile `bareun_rs` (test "test_tokenizer") due to 23 previous errors; 1 warning emitted [INFO] [stderr] warning: build failed, waiting for other jobs to finish... [INFO] [stdout] warning: function `pb_map_to_set` is never used [INFO] [stdout] --> src/custom_dict.rs:39:8 [INFO] [stdout] | [INFO] [stdout] 39 | pub fn pb_map_to_set(ds: &DictSet) -> HashSet { [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: constant `MAX_MESSAGE_LENGTH` is never used [INFO] [stdout] --> src/lang_service_client.rs:12:11 [INFO] [stdout] | [INFO] [stdout] 12 | pub const MAX_MESSAGE_LENGTH: usize = 100 * 1024 * 1024; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `metadata` is never read [INFO] [stdout] --> src/lang_service_client.rs:19:5 [INFO] [stdout] | [INFO] [stdout] 13 | pub struct BareunLanguageServiceClient { [INFO] [stdout] | --------------------------- field in this struct [INFO] [stdout] ... [INFO] [stdout] 19 | metadata: HashMap, [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `phrase` is never read [INFO] [stdout] --> src/tagger.rs:17:5 [INFO] [stdout] | [INFO] [stdout] 12 | pub struct Tagged { [INFO] [stdout] | ------ field in this struct [INFO] [stdout] ... [INFO] [stdout] 17 | phrase: String, [INFO] [stdout] | ^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `phrase` is never read [INFO] [stdout] --> src/tokenizer.rs:18:5 [INFO] [stdout] | [INFO] [stdout] 13 | pub struct Tokenized { [INFO] [stdout] | --------- field in this struct [INFO] [stdout] ... [INFO] [stdout] 18 | phrase: String, [INFO] [stdout] | ^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] running `Command { std: "docker" "inspect" "08240f2ce1cfac83e865ed3b68bf6a2239852106aabf2b50736decfcdfe4b76d", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "08240f2ce1cfac83e865ed3b68bf6a2239852106aabf2b50736decfcdfe4b76d", kill_on_drop: false }` [INFO] [stdout] 08240f2ce1cfac83e865ed3b68bf6a2239852106aabf2b50736decfcdfe4b76d