[INFO] fetching crate bareun_rs 0.1.0... [INFO] checking bareun_rs-0.1.0 against master#350d0ef0ec0493e6d21cfb265cb8211a0e74d766 for pr-145342-1 [INFO] extracting crate bareun_rs 0.1.0 into /workspace/builds/worker-0-tc1/source [INFO] started tweaking crates.io crate bareun_rs 0.1.0 [INFO] finished tweaking crates.io crate bareun_rs 0.1.0 [INFO] tweaked toml for crates.io crate bareun_rs 0.1.0 written to /workspace/builds/worker-0-tc1/source/Cargo.toml [INFO] validating manifest of crates.io crate bareun_rs 0.1.0 on toolchain 350d0ef0ec0493e6d21cfb265cb8211a0e74d766 [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+350d0ef0ec0493e6d21cfb265cb8211a0e74d766" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }` [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+350d0ef0ec0493e6d21cfb265cb8211a0e74d766" "generate-lockfile" "--manifest-path" "Cargo.toml", kill_on_drop: false }` [INFO] [stderr] Updating crates.io index [INFO] [stderr] Locking 140 packages to latest compatible versions [INFO] [stderr] Adding prost v0.12.6 (available: v0.14.1) [INFO] [stderr] Adding tonic v0.11.0 (available: v0.14.1) [INFO] [stderr] Adding tonic-build v0.11.0 (available: v0.14.1) [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+350d0ef0ec0493e6d21cfb265cb8211a0e74d766" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }` [INFO] [stderr] Blocking waiting for file lock on package cache [INFO] [stderr] Downloading crates ... [INFO] [stderr] Downloaded prost-build v0.12.6 [INFO] [stderr] Downloaded tonic-build v0.11.0 [INFO] [stderr] Downloaded tonic v0.11.0 [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-0-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-0-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:7ad1b28ee6f5f7f699f6cf7015098d6ccdd96d6f2d78dd06228f5b4c9faf309c" "/opt/rustwide/cargo-home/bin/cargo" "+350d0ef0ec0493e6d21cfb265cb8211a0e74d766" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }` [INFO] [stdout] 01ec47765426536830ad3d94d271f7069991c4d9ed13959cf8c5925154c41b2a [INFO] running `Command { std: "docker" "start" "-a" "01ec47765426536830ad3d94d271f7069991c4d9ed13959cf8c5925154c41b2a", kill_on_drop: false }` [INFO] running `Command { std: "docker" "inspect" "01ec47765426536830ad3d94d271f7069991c4d9ed13959cf8c5925154c41b2a", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "01ec47765426536830ad3d94d271f7069991c4d9ed13959cf8c5925154c41b2a", kill_on_drop: false }` [INFO] [stdout] 01ec47765426536830ad3d94d271f7069991c4d9ed13959cf8c5925154c41b2a [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-0-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-0-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:7ad1b28ee6f5f7f699f6cf7015098d6ccdd96d6f2d78dd06228f5b4c9faf309c" "/opt/rustwide/cargo-home/bin/cargo" "+350d0ef0ec0493e6d21cfb265cb8211a0e74d766" "check" "--frozen" "--all" "--all-targets" "--message-format=json", kill_on_drop: false }` [INFO] [stdout] a50c0524b763422fae222001526ad778a80448938d51a669c4d8ebc02b7f013a [INFO] running `Command { std: "docker" "start" "-a" "a50c0524b763422fae222001526ad778a80448938d51a669c4d8ebc02b7f013a", kill_on_drop: false }` [INFO] [stderr] Compiling proc-macro2 v1.0.101 [INFO] [stderr] Compiling unicode-ident v1.0.18 [INFO] [stderr] Compiling libc v0.2.175 [INFO] [stderr] Compiling autocfg v1.5.0 [INFO] [stderr] Checking cfg-if v1.0.3 [INFO] [stderr] Checking pin-project-lite v0.2.16 [INFO] [stderr] Checking bytes v1.10.1 [INFO] [stderr] Compiling parking_lot_core v0.9.11 [INFO] [stderr] Checking smallvec v1.15.1 [INFO] [stderr] Checking scopeguard v1.2.0 [INFO] [stderr] Checking futures-core v0.3.31 [INFO] [stderr] Compiling anyhow v1.0.99 [INFO] [stderr] Compiling either v1.15.0 [INFO] [stderr] Compiling rustversion v1.0.22 [INFO] [stderr] Checking itoa v1.0.15 [INFO] [stderr] Checking fnv v1.0.7 [INFO] [stderr] Compiling itertools v0.12.1 [INFO] [stderr] Checking once_cell v1.21.3 [INFO] [stderr] Checking pin-utils v0.1.0 [INFO] [stderr] Compiling lock_api v0.4.13 [INFO] [stderr] Checking futures-task v0.3.31 [INFO] [stderr] Compiling zerocopy v0.8.26 [INFO] [stderr] Checking tracing-core v0.1.34 [INFO] [stderr] Checking futures-util v0.3.31 [INFO] [stderr] Checking http v0.2.12 [INFO] [stderr] Compiling getrandom v0.3.3 [INFO] [stderr] Checking futures-sink v0.3.31 [INFO] [stderr] Compiling rustix v1.0.8 [INFO] [stderr] Compiling quote v1.0.40 [INFO] [stderr] Compiling syn v2.0.106 [INFO] [stderr] Compiling bitflags v2.9.3 [INFO] [stderr] Compiling prettyplease v0.2.37 [INFO] [stderr] Checking mio v1.0.4 [INFO] [stderr] Checking socket2 v0.6.0 [INFO] [stderr] Checking parking_lot v0.12.4 [INFO] [stderr] Checking signal-hook-registry v1.4.6 [INFO] [stderr] Checking getrandom v0.2.16 [INFO] [stderr] Checking hashbrown v0.15.5 [INFO] [stderr] Compiling equivalent v1.0.2 [INFO] [stderr] Compiling linux-raw-sys v0.9.4 [INFO] [stderr] Compiling httparse v1.10.1 [INFO] [stderr] Checking slab v0.4.11 [INFO] [stderr] Checking tower-service v0.3.3 [INFO] [stderr] Compiling regex-syntax v0.8.6 [INFO] [stderr] Checking indexmap v2.11.0 [INFO] [stderr] Checking rand_core v0.6.4 [INFO] [stderr] Checking http-body v0.4.6 [INFO] [stderr] Compiling indexmap v1.9.3 [INFO] [stderr] Compiling fixedbitset v0.4.2 [INFO] [stderr] Checking try-lock v0.2.5 [INFO] [stderr] Compiling serde v1.0.219 [INFO] [stderr] Compiling fastrand v2.3.0 [INFO] [stderr] Checking want v0.3.1 [INFO] [stderr] Checking ppv-lite86 v0.2.21 [INFO] [stderr] Checking socket2 v0.5.10 [INFO] [stderr] Compiling axum-core v0.3.4 [INFO] [stderr] Checking futures-channel v0.3.31 [INFO] [stderr] Checking tower-layer v0.3.3 [INFO] [stderr] Compiling multimap v0.10.1 [INFO] [stderr] Compiling petgraph v0.6.5 [INFO] [stderr] Checking httpdate v1.0.3 [INFO] [stderr] Compiling heck v0.5.0 [INFO] [stderr] Compiling log v0.4.27 [INFO] [stderr] Checking rand_chacha v0.3.1 [INFO] [stderr] Checking hashbrown v0.12.3 [INFO] [stderr] Compiling regex-automata v0.4.10 [INFO] [stderr] Compiling axum v0.6.20 [INFO] [stderr] Checking rand v0.8.5 [INFO] [stderr] Checking memchr v2.7.5 [INFO] [stderr] Checking mime v0.3.17 [INFO] [stderr] Checking percent-encoding v2.3.2 [INFO] [stderr] Checking sync_wrapper v0.1.2 [INFO] [stderr] Checking matchit v0.7.3 [INFO] [stderr] Checking bitflags v1.3.2 [INFO] [stderr] Compiling tempfile v3.21.0 [INFO] [stderr] Compiling serde_json v1.0.143 [INFO] [stderr] Checking base64 v0.21.7 [INFO] [stderr] Checking ryu v1.0.20 [INFO] [stderr] Compiling regex v1.11.2 [INFO] [stderr] Compiling tokio-macros v2.5.0 [INFO] [stderr] Compiling prost-derive v0.12.6 [INFO] [stderr] Compiling tracing-attributes v0.1.30 [INFO] [stderr] Compiling serde_derive v1.0.219 [INFO] [stderr] Compiling pin-project-internal v1.1.10 [INFO] [stderr] Compiling async-trait v0.1.89 [INFO] [stderr] Compiling async-stream-impl v0.3.6 [INFO] [stderr] Checking tokio v1.47.1 [INFO] [stderr] Checking async-stream v0.3.6 [INFO] [stderr] Checking tracing v0.1.41 [INFO] [stderr] Checking pin-project v1.1.10 [INFO] [stderr] Compiling prost v0.12.6 [INFO] [stderr] Compiling prost-types v0.12.6 [INFO] [stderr] Compiling prost-build v0.12.6 [INFO] [stderr] Compiling tonic-build v0.11.0 [INFO] [stderr] Compiling bareun_rs v0.1.0 (/opt/rustwide/workdir) [INFO] [stderr] Checking tokio-util v0.7.16 [INFO] [stderr] Checking tokio-io-timeout v1.2.1 [INFO] [stderr] Checking tokio-stream v0.1.17 [INFO] [stderr] Checking h2 v0.3.27 [INFO] [stderr] Checking tower v0.4.13 [INFO] [stderr] Checking hyper v0.14.32 [INFO] [stderr] Checking hyper-timeout v0.4.1 [INFO] [stderr] Checking tonic v0.11.0 [INFO] [stdout] warning: unused import: `Error` [INFO] [stdout] --> src/lang_service_client.rs:3:42 [INFO] [stdout] | [INFO] [stdout] 3 | use tonic::transport::{channel, Channel, Error}; [INFO] [stdout] | ^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `Tag` [INFO] [stdout] --> src/tagger.rs:3:43 [INFO] [stdout] | [INFO] [stdout] 3 | use crate::bareun::morpheme::{OutOfVocab, Tag}; [INFO] [stdout] | ^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `MAX_MESSAGE_LENGTH` [INFO] [stdout] --> src/tagger.rs:7:63 [INFO] [stdout] | [INFO] [stdout] 7 | use crate::lang_service_client::{BareunLanguageServiceClient, MAX_MESSAGE_LENGTH}; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `tonic::transport::Endpoint` [INFO] [stdout] --> src/tagger.rs:11:5 [INFO] [stdout] | [INFO] [stdout] 11 | use tonic::transport::Endpoint; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `std::collections::HashMap` [INFO] [stdout] --> src/tokenizer.rs:1:5 [INFO] [stdout] | [INFO] [stdout] 1 | use std::collections::HashMap; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused imports: `SegmentToken` and `TextSpan` [INFO] [stdout] --> src/tokenizer.rs:3:47 [INFO] [stdout] | [INFO] [stdout] 3 | use crate::bareun::{Segment, SegmentSentence, SegmentToken, TextSpan, TokenizeResponse}; [INFO] [stdout] | ^^^^^^^^^^^^ ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `MAX_MESSAGE_LENGTH` [INFO] [stdout] --> src/tokenizer.rs:4:63 [INFO] [stdout] | [INFO] [stdout] 4 | use crate::lang_service_client::{BareunLanguageServiceClient, MAX_MESSAGE_LENGTH}; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused imports: `Deserialize` and `Serialize` [INFO] [stdout] --> src/tokenizer.rs:5:13 [INFO] [stdout] | [INFO] [stdout] 5 | use serde::{Deserialize, Serialize}; [INFO] [stdout] | ^^^^^^^^^^^ ^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `serde_json::json` [INFO] [stdout] --> src/tokenizer.rs:7:5 [INFO] [stdout] | [INFO] [stdout] 7 | use serde_json::json; [INFO] [stdout] | ^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `tonic::transport::Endpoint` [INFO] [stdout] --> src/tokenizer.rs:8:5 [INFO] [stdout] | [INFO] [stdout] 8 | use tonic::transport::Endpoint; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `Error` [INFO] [stdout] --> src/lang_service_client.rs:3:42 [INFO] [stdout] | [INFO] [stdout] 3 | use tonic::transport::{channel, Channel, Error}; [INFO] [stdout] | ^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `Tag` [INFO] [stdout] --> src/tagger.rs:3:43 [INFO] [stdout] | [INFO] [stdout] 3 | use crate::bareun::morpheme::{OutOfVocab, Tag}; [INFO] [stdout] | ^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `MAX_MESSAGE_LENGTH` [INFO] [stdout] --> src/tagger.rs:7:63 [INFO] [stdout] | [INFO] [stdout] 7 | use crate::lang_service_client::{BareunLanguageServiceClient, MAX_MESSAGE_LENGTH}; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `tonic::transport::Endpoint` [INFO] [stdout] --> src/tagger.rs:11:5 [INFO] [stdout] | [INFO] [stdout] 11 | use tonic::transport::Endpoint; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `std::collections::HashMap` [INFO] [stdout] --> src/tokenizer.rs:1:5 [INFO] [stdout] | [INFO] [stdout] 1 | use std::collections::HashMap; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused imports: `SegmentToken` and `TextSpan` [INFO] [stdout] --> src/tokenizer.rs:3:47 [INFO] [stdout] | [INFO] [stdout] 3 | use crate::bareun::{Segment, SegmentSentence, SegmentToken, TextSpan, TokenizeResponse}; [INFO] [stdout] | ^^^^^^^^^^^^ ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `MAX_MESSAGE_LENGTH` [INFO] [stdout] --> src/tokenizer.rs:4:63 [INFO] [stdout] | [INFO] [stdout] 4 | use crate::lang_service_client::{BareunLanguageServiceClient, MAX_MESSAGE_LENGTH}; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused imports: `Deserialize` and `Serialize` [INFO] [stdout] --> src/tokenizer.rs:5:13 [INFO] [stdout] | [INFO] [stdout] 5 | use serde::{Deserialize, Serialize}; [INFO] [stdout] | ^^^^^^^^^^^ ^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `serde_json::json` [INFO] [stdout] --> src/tokenizer.rs:7:5 [INFO] [stdout] | [INFO] [stdout] 7 | use serde_json::json; [INFO] [stdout] | ^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `tonic::transport::Endpoint` [INFO] [stdout] --> src/tokenizer.rs:8:5 [INFO] [stdout] | [INFO] [stdout] 8 | use tonic::transport::Endpoint; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: function `pb_map_to_set` is never used [INFO] [stdout] --> src/custom_dict.rs:39:8 [INFO] [stdout] | [INFO] [stdout] 39 | pub fn pb_map_to_set(ds: &DictSet) -> HashSet { [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(dead_code)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: constant `MAX_MESSAGE_LENGTH` is never used [INFO] [stdout] --> src/lang_service_client.rs:12:11 [INFO] [stdout] | [INFO] [stdout] 12 | pub const MAX_MESSAGE_LENGTH: usize = 100 * 1024 * 1024; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `metadata` is never read [INFO] [stdout] --> src/lang_service_client.rs:19:5 [INFO] [stdout] | [INFO] [stdout] 13 | pub struct BareunLanguageServiceClient { [INFO] [stdout] | --------------------------- field in this struct [INFO] [stdout] ... [INFO] [stdout] 19 | metadata: HashMap, [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `phrase` is never read [INFO] [stdout] --> src/tagger.rs:17:5 [INFO] [stdout] | [INFO] [stdout] 12 | pub struct Tagged { [INFO] [stdout] | ------ field in this struct [INFO] [stdout] ... [INFO] [stdout] 17 | phrase: String, [INFO] [stdout] | ^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `phrase` is never read [INFO] [stdout] --> src/tokenizer.rs:18:5 [INFO] [stdout] | [INFO] [stdout] 13 | pub struct Tokenized { [INFO] [stdout] | --------- field in this struct [INFO] [stdout] ... [INFO] [stdout] 18 | phrase: String, [INFO] [stdout] | ^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `bareun_rs::Tokenized` [INFO] [stdout] --> tests/test_tokenizer.rs:3:9 [INFO] [stdout] | [INFO] [stdout] 3 | use bareun_rs::Tokenized; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0061]: this method takes 4 arguments but 2 arguments were supplied [INFO] [stdout] --> tests/test_tokenizer.rs:16:29 [INFO] [stdout] | [INFO] [stdout] 16 | tokenizer.await.seg(TEST_STR, false), [INFO] [stdout] | ^^^----------------- two arguments of type `bool` and `bool` are missing [INFO] [stdout] | [INFO] [stdout] note: method defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:381:18 [INFO] [stdout] | [INFO] [stdout] 381 | pub async fn seg( [INFO] [stdout] | ^^^ [INFO] [stdout] help: provide the arguments [INFO] [stdout] | [INFO] [stdout] 16 | tokenizer.await.seg(TEST_STR, false, /* bool */, /* bool */), [INFO] [stdout] | ++++++++++++++++++++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0369]: binary operation `==` cannot be applied to type `impl Future` [INFO] [stdout] --> tests/test_tokenizer.rs:15:9 [INFO] [stdout] | [INFO] [stdout] 15 | / assert_eq!( [INFO] [stdout] 16 | | tokenizer.await.seg(TEST_STR, false), [INFO] [stdout] 17 | | vec![ [INFO] [stdout] 18 | | vec!["오늘".to_string(), "은".to_string()], [INFO] [stdout] ... | [INFO] [stdout] 29 | | ); [INFO] [stdout] | | ^ [INFO] [stdout] | | | [INFO] [stdout] | |_________impl Future [INFO] [stdout] | Vec> [INFO] [stdout] | [INFO] [stdout] = note: this error originates in the macro `assert_eq` (in Nightly builds, run with -Z macro-backtrace for more info) [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0277]: `impl Future` doesn't implement `Debug` [INFO] [stdout] --> tests/test_tokenizer.rs:15:9 [INFO] [stdout] | [INFO] [stdout] 15 | / assert_eq!( [INFO] [stdout] 16 | | tokenizer.await.seg(TEST_STR, false), [INFO] [stdout] 17 | | vec![ [INFO] [stdout] 18 | | vec!["오늘".to_string(), "은".to_string()], [INFO] [stdout] ... | [INFO] [stdout] 29 | | ); [INFO] [stdout] | |_________^ the trait `Debug` is not implemented for `impl Future` [INFO] [stdout] | [INFO] [stdout] = note: this error originates in the macro `assert_eq` (in Nightly builds, run with -Z macro-backtrace for more info) [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:34:93 [INFO] [stdout] | [INFO] [stdout] 34 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 34 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `seg` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:36:23 [INFO] [stdout] | [INFO] [stdout] 36 | tokenizer.seg(TEST_STR, true, true), [INFO] [stdout] | ^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 36 | tokenizer.await.seg(TEST_STR, true, true), [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:54:93 [INFO] [stdout] | [INFO] [stdout] 54 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 54 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `seg` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:55:30 [INFO] [stdout] | [INFO] [stdout] 55 | let temp = tokenizer.seg(TEST_STR, true); [INFO] [stdout] | ^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 55 | let temp = tokenizer.await.seg(TEST_STR, true); [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:76:93 [INFO] [stdout] | [INFO] [stdout] 76 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 76 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `seg` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:78:23 [INFO] [stdout] | [INFO] [stdout] 78 | tokenizer.seg(TEST_STR), [INFO] [stdout] | ^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 78 | tokenizer.await.seg(TEST_STR), [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:96:93 [INFO] [stdout] | [INFO] [stdout] 96 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 96 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `nouns` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:98:23 [INFO] [stdout] | [INFO] [stdout] 98 | tokenizer.nouns(TEST_STR), [INFO] [stdout] | ^^^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 98 | tokenizer.await.nouns(TEST_STR), [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:105:93 [INFO] [stdout] | [INFO] [stdout] 105 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 105 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `tokenize` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:106:35 [INFO] [stdout] | [INFO] [stdout] 106 | let tokenized = tokenizer.tokenize(TEST_STR); [INFO] [stdout] | ^^^^^^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 106 | let tokenized = tokenizer.await.tokenize(TEST_STR); [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:115:93 [INFO] [stdout] | [INFO] [stdout] 115 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 115 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `tokenize` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:116:35 [INFO] [stdout] | [INFO] [stdout] 116 | let tokenized = tokenizer.tokenize(TEST_STR); [INFO] [stdout] | ^^^^^^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 116 | let tokenized = tokenizer.await.tokenize(TEST_STR); [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:125:93 [INFO] [stdout] | [INFO] [stdout] 125 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 125 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `tokenize` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:126:35 [INFO] [stdout] | [INFO] [stdout] 126 | let tokenized = tokenizer.tokenize(TEST_STR); [INFO] [stdout] | ^^^^^^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 126 | let tokenized = tokenizer.await.tokenize(TEST_STR); [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:132:93 [INFO] [stdout] | [INFO] [stdout] 132 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 132 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `tokenize` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:133:35 [INFO] [stdout] | [INFO] [stdout] 133 | let tokenized = tokenizer.tokenize(TEST_STR); [INFO] [stdout] | ^^^^^^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 133 | let tokenized = tokenizer.await.tokenize(TEST_STR); [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:139:93 [INFO] [stdout] | [INFO] [stdout] 139 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 139 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `tokenize` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:140:35 [INFO] [stdout] | [INFO] [stdout] 140 | let tokenized = tokenizer.tokenize(TEST_STR); [INFO] [stdout] | ^^^^^^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 140 | let tokenized = tokenizer.await.tokenize(TEST_STR); [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0308]: mismatched types [INFO] [stdout] --> tests/test_tokenizer.rs:149:93 [INFO] [stdout] | [INFO] [stdout] 149 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", 5757); [INFO] [stdout] | -------------- arguments to this function are incorrect ^^^^ expected `Option`, found integer [INFO] [stdout] | [INFO] [stdout] = note: expected enum `Option` [INFO] [stdout] found type `{integer}` [INFO] [stdout] note: associated function defined here [INFO] [stdout] --> /opt/rustwide/workdir/src/tokenizer.rs:322:18 [INFO] [stdout] | [INFO] [stdout] 322 | pub async fn new(apikey: &str, host: &str, port: Option) -> Self { [INFO] [stdout] | ^^^ [INFO] [stdout] help: try wrapping the expression in `Some` [INFO] [stdout] | [INFO] [stdout] 149 | let tokenizer = Tokenizer::new("koba-AH6GVAA-5AKEGUY-QN3JJEI-3NRX6JY", "127.0.0.1", Some(5757)); [INFO] [stdout] | +++++ + [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] error[E0599]: no method named `tokenize` found for opaque type `impl Future` in the current scope [INFO] [stdout] --> tests/test_tokenizer.rs:150:35 [INFO] [stdout] | [INFO] [stdout] 150 | let tokenized = tokenizer.tokenize(TEST_STR); [INFO] [stdout] | ^^^^^^^^ method not found in `impl Future` [INFO] [stdout] | [INFO] [stdout] help: consider `await`ing on the `Future` and calling the method on its `Output` [INFO] [stdout] | [INFO] [stdout] 150 | let tokenized = tokenizer.await.tokenize(TEST_STR); [INFO] [stdout] | ++++++ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] Some errors have detailed explanations: E0061, E0277, E0308, E0369, E0599. [INFO] [stdout] [INFO] [stdout] For more information about an error, try `rustc --explain E0061`. [INFO] [stdout] [INFO] [stdout] warning: function `pb_map_to_set` is never used [INFO] [stdout] --> src/custom_dict.rs:39:8 [INFO] [stdout] | [INFO] [stdout] 39 | pub fn pb_map_to_set(ds: &DictSet) -> HashSet { [INFO] [stdout] | ^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(dead_code)]` on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: constant `MAX_MESSAGE_LENGTH` is never used [INFO] [stdout] --> src/lang_service_client.rs:12:11 [INFO] [stdout] | [INFO] [stdout] 12 | pub const MAX_MESSAGE_LENGTH: usize = 100 * 1024 * 1024; [INFO] [stdout] | ^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `metadata` is never read [INFO] [stdout] --> src/lang_service_client.rs:19:5 [INFO] [stdout] | [INFO] [stdout] 13 | pub struct BareunLanguageServiceClient { [INFO] [stdout] | --------------------------- field in this struct [INFO] [stdout] ... [INFO] [stdout] 19 | metadata: HashMap, [INFO] [stdout] | ^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `phrase` is never read [INFO] [stdout] --> src/tagger.rs:17:5 [INFO] [stdout] | [INFO] [stdout] 12 | pub struct Tagged { [INFO] [stdout] | ------ field in this struct [INFO] [stdout] ... [INFO] [stdout] 17 | phrase: String, [INFO] [stdout] | ^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `phrase` is never read [INFO] [stdout] --> src/tokenizer.rs:18:5 [INFO] [stdout] | [INFO] [stdout] 13 | pub struct Tokenized { [INFO] [stdout] | --------- field in this struct [INFO] [stdout] ... [INFO] [stdout] 18 | phrase: String, [INFO] [stdout] | ^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] error: could not compile `bareun_rs` (test "test_tokenizer") due to 23 previous errors; 1 warning emitted [INFO] [stderr] warning: build failed, waiting for other jobs to finish... [INFO] running `Command { std: "docker" "inspect" "a50c0524b763422fae222001526ad778a80448938d51a669c4d8ebc02b7f013a", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "a50c0524b763422fae222001526ad778a80448938d51a669c4d8ebc02b7f013a", kill_on_drop: false }` [INFO] [stdout] a50c0524b763422fae222001526ad778a80448938d51a669c4d8ebc02b7f013a