[INFO] fetching crate llama_cpp_rs 0.3.0... [INFO] checking llama_cpp_rs-0.3.0 against try#a4980311fb7bb9e7893708e6bd3fbbfb2819fd3d for pr-145342-1 [INFO] extracting crate llama_cpp_rs 0.3.0 into /workspace/builds/worker-3-tc2/source [INFO] started tweaking crates.io crate llama_cpp_rs 0.3.0 [INFO] finished tweaking crates.io crate llama_cpp_rs 0.3.0 [INFO] tweaked toml for crates.io crate llama_cpp_rs 0.3.0 written to /workspace/builds/worker-3-tc2/source/Cargo.toml [INFO] validating manifest of crates.io crate llama_cpp_rs 0.3.0 on toolchain a4980311fb7bb9e7893708e6bd3fbbfb2819fd3d [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+a4980311fb7bb9e7893708e6bd3fbbfb2819fd3d" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }` [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+a4980311fb7bb9e7893708e6bd3fbbfb2819fd3d" "generate-lockfile" "--manifest-path" "Cargo.toml", kill_on_drop: false }` [INFO] [stderr] Updating crates.io index [INFO] [stderr] Locking 55 packages to latest compatible versions [INFO] [stderr] Adding bindgen v0.66.1 (available: v0.72.0) [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+a4980311fb7bb9e7893708e6bd3fbbfb2819fd3d" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }` [INFO] [stderr] Blocking waiting for file lock on package cache [INFO] [stderr] Blocking waiting for file lock on package cache [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-3-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-3-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:7ad1b28ee6f5f7f699f6cf7015098d6ccdd96d6f2d78dd06228f5b4c9faf309c" "/opt/rustwide/cargo-home/bin/cargo" "+a4980311fb7bb9e7893708e6bd3fbbfb2819fd3d" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }` [INFO] [stdout] a2d85f6c478cb8beee77a38fe454c7e8ec64c5285c4191341047fa187c7c649e [INFO] running `Command { std: "docker" "start" "-a" "a2d85f6c478cb8beee77a38fe454c7e8ec64c5285c4191341047fa187c7c649e", kill_on_drop: false }` [INFO] running `Command { std: "docker" "inspect" "a2d85f6c478cb8beee77a38fe454c7e8ec64c5285c4191341047fa187c7c649e", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "a2d85f6c478cb8beee77a38fe454c7e8ec64c5285c4191341047fa187c7c649e", kill_on_drop: false }` [INFO] [stdout] a2d85f6c478cb8beee77a38fe454c7e8ec64c5285c4191341047fa187c7c649e [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-3-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-3-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:7ad1b28ee6f5f7f699f6cf7015098d6ccdd96d6f2d78dd06228f5b4c9faf309c" "/opt/rustwide/cargo-home/bin/cargo" "+a4980311fb7bb9e7893708e6bd3fbbfb2819fd3d" "check" "--frozen" "--all" "--all-targets" "--message-format=json", kill_on_drop: false }` [INFO] [stdout] 06c50779f00dbda1a3312e1a88d5c4a5bfaad4f8a668164588b4068290f336fe [INFO] running `Command { std: "docker" "start" "-a" "06c50779f00dbda1a3312e1a88d5c4a5bfaad4f8a668164588b4068290f336fe", kill_on_drop: false }` [INFO] [stderr] Compiling proc-macro2 v1.0.101 [INFO] [stderr] Compiling unicode-ident v1.0.18 [INFO] [stderr] Compiling glob v0.3.3 [INFO] [stderr] Compiling libc v0.2.175 [INFO] [stderr] Compiling prettyplease v0.2.37 [INFO] [stderr] Compiling rustix v0.38.44 [INFO] [stderr] Compiling memchr v2.7.5 [INFO] [stderr] Compiling bitflags v2.9.3 [INFO] [stderr] Compiling cfg-if v1.0.3 [INFO] [stderr] Compiling minimal-lexical v0.2.1 [INFO] [stderr] Compiling linux-raw-sys v0.4.15 [INFO] [stderr] Compiling regex-syntax v0.8.6 [INFO] [stderr] Compiling shlex v1.3.0 [INFO] [stderr] Compiling either v1.15.0 [INFO] [stderr] Compiling home v0.5.11 [INFO] [stderr] Compiling bindgen v0.66.1 [INFO] [stderr] Compiling peeking_take_while v0.1.2 [INFO] [stderr] Compiling libloading v0.8.8 [INFO] [stderr] Compiling lazycell v1.3.0 [INFO] [stderr] Compiling log v0.4.27 [INFO] [stderr] Compiling lazy_static v1.5.0 [INFO] [stderr] Compiling rustc-hash v1.1.0 [INFO] [stderr] Compiling cc v1.2.34 [INFO] [stderr] Compiling clang-sys v1.8.1 [INFO] [stderr] Compiling nom v7.1.3 [INFO] [stderr] Compiling quote v1.0.40 [INFO] [stderr] Compiling syn v2.0.106 [INFO] [stderr] Compiling regex-automata v0.4.10 [INFO] [stderr] Compiling cexpr v0.6.0 [INFO] [stderr] Compiling which v4.4.2 [INFO] [stderr] Compiling regex v1.11.2 [INFO] [stderr] Compiling llama_cpp_rs v0.3.0 (/opt/rustwide/workdir) [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/ggml.c:19672:13: warning: 'ggml_opt_get_grad' defined but not used [-Wunused-function] [INFO] [stderr] warning: llama_cpp_rs@0.3.0: 19672 | static void ggml_opt_get_grad(int np, struct ggml_tensor * const ps[], float * g) { [INFO] [stderr] warning: llama_cpp_rs@0.3.0: | ^~~~~~~~~~~~~~~~~ [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/k_quants.c:205:14: warning: 'make_qkx1_quants' defined but not used [-Wunused-function] [INFO] [stderr] warning: llama_cpp_rs@0.3.0: 205 | static float make_qkx1_quants(int n, int nmax, const float * restrict x, uint8_t * restrict L, float * restrict the_min, [INFO] [stderr] warning: llama_cpp_rs@0.3.0: | ^~~~~~~~~~~~~~~~ [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'int get_embeddings(void*, void*, float*)': [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:80:23: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations] [INFO] [stderr] warning: llama_cpp_rs@0.3.0: 80 | if (llama_eval(ctx, embd_inp.data(), embd_inp.size(), n_past)) [INFO] [stderr] warning: llama_cpp_rs@0.3.0: | ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ [INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5, [INFO] [stderr] warning: llama_cpp_rs@0.3.0: from ./binding.cpp:1: [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:423:30: note: declared here [INFO] [stderr] warning: llama_cpp_rs@0.3.0: 423 | LLAMA_API DEPRECATED(int llama_eval( [INFO] [stderr] warning: llama_cpp_rs@0.3.0: | ^~~~~~~~~~ [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED' [INFO] [stderr] warning: llama_cpp_rs@0.3.0: 31 | # define DEPRECATED(func, hint) func __attribute__((deprecated(hint))) [INFO] [stderr] warning: llama_cpp_rs@0.3.0: | ^~~~ [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'int eval(void*, void*, char*)': [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:138:22: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations] [INFO] [stderr] warning: llama_cpp_rs@0.3.0: 138 | return llama_eval(ctx, tokens.data(), n_prompt_tokens, n_past); [INFO] [stderr] warning: llama_cpp_rs@0.3.0: | ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:423:30: note: declared here [INFO] [stderr] warning: llama_cpp_rs@0.3.0: 423 | LLAMA_API DEPRECATED(int llama_eval( [INFO] [stderr] warning: llama_cpp_rs@0.3.0: | ^~~~~~~~~~ [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED' [INFO] [stderr] warning: llama_cpp_rs@0.3.0: 31 | # define DEPRECATED(func, hint) func __attribute__((deprecated(hint))) [INFO] [stderr] warning: llama_cpp_rs@0.3.0: | ^~~~ [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'int llama_predict(void*, void*, char*, bool)': [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:282:19: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations] [INFO] [stderr] warning: llama_cpp_rs@0.3.0: 282 | llama_eval(ctx, tmp, 1, 0); [INFO] [stderr] warning: llama_cpp_rs@0.3.0: | ~~~~~~~~~~^~~~~~~~~~~~~~~~ [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:423:30: note: declared here [INFO] [stderr] warning: llama_cpp_rs@0.3.0: 423 | LLAMA_API DEPRECATED(int llama_eval( [INFO] [stderr] warning: llama_cpp_rs@0.3.0: | ^~~~~~~~~~ [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED' [INFO] [stderr] warning: llama_cpp_rs@0.3.0: 31 | # define DEPRECATED(func, hint) func __attribute__((deprecated(hint))) [INFO] [stderr] warning: llama_cpp_rs@0.3.0: | ^~~~ [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:353:31: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations] [INFO] [stderr] warning: llama_cpp_rs@0.3.0: 353 | if (llama_eval(ctx, &embd[i], n_eval, n_past)) [INFO] [stderr] warning: llama_cpp_rs@0.3.0: | ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:423:30: note: declared here [INFO] [stderr] warning: llama_cpp_rs@0.3.0: 423 | LLAMA_API DEPRECATED(int llama_eval( [INFO] [stderr] warning: llama_cpp_rs@0.3.0: | ^~~~~~~~~~ [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED' [INFO] [stderr] warning: llama_cpp_rs@0.3.0: 31 | # define DEPRECATED(func, hint) func __attribute__((deprecated(hint))) [INFO] [stderr] warning: llama_cpp_rs@0.3.0: | ^~~~ [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:440:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations] [INFO] [stderr] warning: llama_cpp_rs@0.3.0: 440 | llama_sample_temperature(ctx, &candidates_p, temp); [INFO] [stderr] warning: llama_cpp_rs@0.3.0: | ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~ [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:621:31: note: declared here [INFO] [stderr] warning: llama_cpp_rs@0.3.0: 621 | LLAMA_API DEPRECATED(void llama_sample_temperature( [INFO] [stderr] warning: llama_cpp_rs@0.3.0: | ^~~~~~~~~~~~~~~~~~~~~~~~ [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED' [INFO] [stderr] warning: llama_cpp_rs@0.3.0: 31 | # define DEPRECATED(func, hint) func __attribute__((deprecated(hint))) [INFO] [stderr] warning: llama_cpp_rs@0.3.0: | ^~~~ [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:446:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations] [INFO] [stderr] warning: llama_cpp_rs@0.3.0: 446 | llama_sample_temperature(ctx, &candidates_p, temp); [INFO] [stderr] warning: llama_cpp_rs@0.3.0: | ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~ [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:621:31: note: declared here [INFO] [stderr] warning: llama_cpp_rs@0.3.0: 621 | LLAMA_API DEPRECATED(void llama_sample_temperature( [INFO] [stderr] warning: llama_cpp_rs@0.3.0: | ^~~~~~~~~~~~~~~~~~~~~~~~ [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED' [INFO] [stderr] warning: llama_cpp_rs@0.3.0: 31 | # define DEPRECATED(func, hint) func __attribute__((deprecated(hint))) [INFO] [stderr] warning: llama_cpp_rs@0.3.0: | ^~~~ [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:456:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations] [INFO] [stderr] warning: llama_cpp_rs@0.3.0: 456 | llama_sample_temperature(ctx, &candidates_p, temp); [INFO] [stderr] warning: llama_cpp_rs@0.3.0: | ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~ [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:621:31: note: declared here [INFO] [stderr] warning: llama_cpp_rs@0.3.0: 621 | LLAMA_API DEPRECATED(void llama_sample_temperature( [INFO] [stderr] warning: llama_cpp_rs@0.3.0: | ^~~~~~~~~~~~~~~~~~~~~~~~ [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED' [INFO] [stderr] warning: llama_cpp_rs@0.3.0: 31 | # define DEPRECATED(func, hint) func __attribute__((deprecated(hint))) [INFO] [stderr] warning: llama_cpp_rs@0.3.0: | ^~~~ [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:475:42: warning: cast from type 'const char*' to type 'char*' casts away qualifiers [-Wcast-qual] [INFO] [stderr] warning: llama_cpp_rs@0.3.0: 475 | if (!tokenCallback(state_pr, (char*)token_str.c_str())) [INFO] [stderr] warning: llama_cpp_rs@0.3.0: | ^~~~~~~~~~~~~~~~~~~~~~~~ [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'void* llama_allocate_params(const char*, int, int, int, int, float, float, float, int, bool, bool, int, int, const char**, int, float, float, float, float, int, float, float, bool, const char*, const char*, bool, bool, bool, const char*, const char*, bool)': [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:629:100: warning: unused parameter 'ignore_eos' [-Wunused-parameter] [INFO] [stderr] warning: llama_cpp_rs@0.3.0: 629 | float top_p, float temp, float repeat_penalty, int repeat_last_n, bool ignore_eos, bool memory_f16, int n_batch, int n_keep, const char **antiprompt, int antiprompt_count, [INFO] [stderr] warning: llama_cpp_rs@0.3.0: | ~~~~~^~~~~~~~~~ [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'void* load_model(const char*, int, int, bool, bool, bool, bool, bool, bool, int, int, const char*, const char*, bool)': [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:708:122: warning: unused parameter 'low_vram' [-Wunused-parameter] [INFO] [stderr] warning: llama_cpp_rs@0.3.0: 708 | void *load_model(const char *fname, int n_ctx, int n_seed, bool memory_f16, bool mlock, bool embeddings, bool mmap, bool low_vram, bool vocab_only, int n_gpu_layers, int n_batch, const char *maingpu, const char *tensorsplit, bool numa) [INFO] [stderr] warning: llama_cpp_rs@0.3.0: | ~~~~~^~~~~~~~ [INFO] [stderr] warning: llama_cpp_rs@0.3.0: ar: /opt/rustwide/target/debug/build/llama_cpp_rs-f2e7c9ea871f58df/out/llama.cpp/ggml.o: No such file or directory [INFO] [stderr] error: failed to run custom build command for `llama_cpp_rs v0.3.0 (/opt/rustwide/workdir)` [INFO] [stderr] note: To improve backtraces for build dependencies, set the CARGO_PROFILE_DEV_BUILD_OVERRIDE_DEBUG=true environment variable to enable debug information generation. [INFO] [stderr] [INFO] [stderr] Caused by: [INFO] [stderr] process didn't exit successfully: `/opt/rustwide/target/debug/build/llama_cpp_rs-9479c655e9f8bc46/build-script-build` (exit status: 1) [INFO] [stderr] --- stdout [INFO] [stderr] cargo:rerun-if-env-changed=TARGET [INFO] [stderr] cargo:rerun-if-env-changed=BINDGEN_EXTRA_CLANG_ARGS_x86_64-unknown-linux-gnu [INFO] [stderr] cargo:rerun-if-env-changed=BINDGEN_EXTRA_CLANG_ARGS_x86_64_unknown_linux_gnu [INFO] [stderr] cargo:rerun-if-env-changed=BINDGEN_EXTRA_CLANG_ARGS [INFO] [stderr] cargo:rerun-if-changed=/usr/lib/llvm-18/lib/clang/18/include/stdbool.h [INFO] [stderr] OUT_DIR = Some(/opt/rustwide/target/debug/build/llama_cpp_rs-f2e7c9ea871f58df/out) [INFO] [stderr] OPT_LEVEL = Some(0) [INFO] [stderr] TARGET = Some(x86_64-unknown-linux-gnu) [INFO] [stderr] HOST = Some(x86_64-unknown-linux-gnu) [INFO] [stderr] cargo:rerun-if-env-changed=CC_x86_64-unknown-linux-gnu [INFO] [stderr] CC_x86_64-unknown-linux-gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=CC_x86_64_unknown_linux_gnu [INFO] [stderr] CC_x86_64_unknown_linux_gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=HOST_CC [INFO] [stderr] HOST_CC = None [INFO] [stderr] cargo:rerun-if-env-changed=CC [INFO] [stderr] CC = None [INFO] [stderr] cargo:rerun-if-env-changed=CC_ENABLE_DEBUG_OUTPUT [INFO] [stderr] RUSTC_WRAPPER = None [INFO] [stderr] cargo:rerun-if-env-changed=CRATE_CC_NO_DEFAULTS [INFO] [stderr] CRATE_CC_NO_DEFAULTS = None [INFO] [stderr] DEBUG = Some(true) [INFO] [stderr] CARGO_CFG_TARGET_FEATURE = Some(fxsr,sse,sse2,x87) [INFO] [stderr] cargo:rerun-if-env-changed=CFLAGS [INFO] [stderr] CFLAGS = None [INFO] [stderr] cargo:rerun-if-env-changed=HOST_CFLAGS [INFO] [stderr] HOST_CFLAGS = None [INFO] [stderr] cargo:rerun-if-env-changed=CFLAGS_x86_64_unknown_linux_gnu [INFO] [stderr] CFLAGS_x86_64_unknown_linux_gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=CFLAGS_x86_64-unknown-linux-gnu [INFO] [stderr] CFLAGS_x86_64-unknown-linux-gnu = None [INFO] [stderr] CARGO_ENCODED_RUSTFLAGS = Some(--cap-lints=forbid) [INFO] [stderr] cargo:warning=./llama.cpp/ggml.c:19672:13: warning: 'ggml_opt_get_grad' defined but not used [-Wunused-function] [INFO] [stderr] cargo:warning=19672 | static void ggml_opt_get_grad(int np, struct ggml_tensor * const ps[], float * g) { [INFO] [stderr] cargo:warning= | ^~~~~~~~~~~~~~~~~ [INFO] [stderr] cargo:warning=./llama.cpp/k_quants.c:205:14: warning: 'make_qkx1_quants' defined but not used [-Wunused-function] [INFO] [stderr] cargo:warning= 205 | static float make_qkx1_quants(int n, int nmax, const float * restrict x, uint8_t * restrict L, float * restrict the_min, [INFO] [stderr] cargo:warning= | ^~~~~~~~~~~~~~~~ [INFO] [stderr] cargo:rerun-if-env-changed=AR_x86_64-unknown-linux-gnu [INFO] [stderr] AR_x86_64-unknown-linux-gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=AR_x86_64_unknown_linux_gnu [INFO] [stderr] AR_x86_64_unknown_linux_gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=HOST_AR [INFO] [stderr] HOST_AR = None [INFO] [stderr] cargo:rerun-if-env-changed=AR [INFO] [stderr] AR = None [INFO] [stderr] cargo:rerun-if-env-changed=ARFLAGS [INFO] [stderr] ARFLAGS = None [INFO] [stderr] cargo:rerun-if-env-changed=HOST_ARFLAGS [INFO] [stderr] HOST_ARFLAGS = None [INFO] [stderr] cargo:rerun-if-env-changed=ARFLAGS_x86_64_unknown_linux_gnu [INFO] [stderr] ARFLAGS_x86_64_unknown_linux_gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=ARFLAGS_x86_64-unknown-linux-gnu [INFO] [stderr] ARFLAGS_x86_64-unknown-linux-gnu = None [INFO] [stderr] cargo:rustc-link-lib=static=ggml [INFO] [stderr] cargo:rustc-link-search=native=/opt/rustwide/target/debug/build/llama_cpp_rs-f2e7c9ea871f58df/out [INFO] [stderr] OUT_DIR = Some(/opt/rustwide/target/debug/build/llama_cpp_rs-f2e7c9ea871f58df/out) [INFO] [stderr] OPT_LEVEL = Some(0) [INFO] [stderr] TARGET = Some(x86_64-unknown-linux-gnu) [INFO] [stderr] HOST = Some(x86_64-unknown-linux-gnu) [INFO] [stderr] cargo:rerun-if-env-changed=CXX_x86_64-unknown-linux-gnu [INFO] [stderr] CXX_x86_64-unknown-linux-gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=CXX_x86_64_unknown_linux_gnu [INFO] [stderr] CXX_x86_64_unknown_linux_gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=HOST_CXX [INFO] [stderr] HOST_CXX = None [INFO] [stderr] cargo:rerun-if-env-changed=CXX [INFO] [stderr] CXX = None [INFO] [stderr] cargo:rerun-if-env-changed=CC_ENABLE_DEBUG_OUTPUT [INFO] [stderr] RUSTC_WRAPPER = None [INFO] [stderr] cargo:rerun-if-env-changed=CRATE_CC_NO_DEFAULTS [INFO] [stderr] CRATE_CC_NO_DEFAULTS = None [INFO] [stderr] DEBUG = Some(true) [INFO] [stderr] CARGO_CFG_TARGET_FEATURE = Some(fxsr,sse,sse2,x87) [INFO] [stderr] cargo:rerun-if-env-changed=CXXFLAGS [INFO] [stderr] CXXFLAGS = None [INFO] [stderr] cargo:rerun-if-env-changed=HOST_CXXFLAGS [INFO] [stderr] HOST_CXXFLAGS = None [INFO] [stderr] cargo:rerun-if-env-changed=CXXFLAGS_x86_64_unknown_linux_gnu [INFO] [stderr] CXXFLAGS_x86_64_unknown_linux_gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=CXXFLAGS_x86_64-unknown-linux-gnu [INFO] [stderr] CXXFLAGS_x86_64-unknown-linux-gnu = None [INFO] [stderr] CARGO_ENCODED_RUSTFLAGS = Some(--cap-lints=forbid) [INFO] [stderr] cargo:warning=./binding.cpp: In function 'int get_embeddings(void*, void*, float*)': [INFO] [stderr] cargo:warning=./binding.cpp:80:23: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations] [INFO] [stderr] cargo:warning= 80 | if (llama_eval(ctx, embd_inp.data(), embd_inp.size(), n_past)) [INFO] [stderr] cargo:warning= | ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ [INFO] [stderr] cargo:warning=In file included from ./llama.cpp/common/common.h:5, [INFO] [stderr] cargo:warning= from ./binding.cpp:1: [INFO] [stderr] cargo:warning=./llama.cpp/llama.h:423:30: note: declared here [INFO] [stderr] cargo:warning= 423 | LLAMA_API DEPRECATED(int llama_eval( [INFO] [stderr] cargo:warning= | ^~~~~~~~~~ [INFO] [stderr] cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED' [INFO] [stderr] cargo:warning= 31 | # define DEPRECATED(func, hint) func __attribute__((deprecated(hint))) [INFO] [stderr] cargo:warning= | ^~~~ [INFO] [stderr] cargo:warning=./binding.cpp: In function 'int eval(void*, void*, char*)': [INFO] [stderr] cargo:warning=./binding.cpp:138:22: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations] [INFO] [stderr] cargo:warning= 138 | return llama_eval(ctx, tokens.data(), n_prompt_tokens, n_past); [INFO] [stderr] cargo:warning= | ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ [INFO] [stderr] cargo:warning=./llama.cpp/llama.h:423:30: note: declared here [INFO] [stderr] cargo:warning= 423 | LLAMA_API DEPRECATED(int llama_eval( [INFO] [stderr] cargo:warning= | ^~~~~~~~~~ [INFO] [stderr] cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED' [INFO] [stderr] cargo:warning= 31 | # define DEPRECATED(func, hint) func __attribute__((deprecated(hint))) [INFO] [stderr] cargo:warning= | ^~~~ [INFO] [stderr] cargo:warning=./binding.cpp: In function 'int llama_predict(void*, void*, char*, bool)': [INFO] [stderr] cargo:warning=./binding.cpp:282:19: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations] [INFO] [stderr] cargo:warning= 282 | llama_eval(ctx, tmp, 1, 0); [INFO] [stderr] cargo:warning= | ~~~~~~~~~~^~~~~~~~~~~~~~~~ [INFO] [stderr] cargo:warning=./llama.cpp/llama.h:423:30: note: declared here [INFO] [stderr] cargo:warning= 423 | LLAMA_API DEPRECATED(int llama_eval( [INFO] [stderr] cargo:warning= | ^~~~~~~~~~ [INFO] [stderr] cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED' [INFO] [stderr] cargo:warning= 31 | # define DEPRECATED(func, hint) func __attribute__((deprecated(hint))) [INFO] [stderr] cargo:warning= | ^~~~ [INFO] [stderr] cargo:warning=./binding.cpp:353:31: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations] [INFO] [stderr] cargo:warning= 353 | if (llama_eval(ctx, &embd[i], n_eval, n_past)) [INFO] [stderr] cargo:warning= | ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ [INFO] [stderr] cargo:warning=./llama.cpp/llama.h:423:30: note: declared here [INFO] [stderr] cargo:warning= 423 | LLAMA_API DEPRECATED(int llama_eval( [INFO] [stderr] cargo:warning= | ^~~~~~~~~~ [INFO] [stderr] cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED' [INFO] [stderr] cargo:warning= 31 | # define DEPRECATED(func, hint) func __attribute__((deprecated(hint))) [INFO] [stderr] cargo:warning= | ^~~~ [INFO] [stderr] cargo:warning=./binding.cpp:440:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations] [INFO] [stderr] cargo:warning= 440 | llama_sample_temperature(ctx, &candidates_p, temp); [INFO] [stderr] cargo:warning= | ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~ [INFO] [stderr] cargo:warning=./llama.cpp/llama.h:621:31: note: declared here [INFO] [stderr] cargo:warning= 621 | LLAMA_API DEPRECATED(void llama_sample_temperature( [INFO] [stderr] cargo:warning= | ^~~~~~~~~~~~~~~~~~~~~~~~ [INFO] [stderr] cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED' [INFO] [stderr] cargo:warning= 31 | # define DEPRECATED(func, hint) func __attribute__((deprecated(hint))) [INFO] [stderr] cargo:warning= | ^~~~ [INFO] [stderr] cargo:warning=./binding.cpp:446:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations] [INFO] [stderr] cargo:warning= 446 | llama_sample_temperature(ctx, &candidates_p, temp); [INFO] [stderr] cargo:warning= | ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~ [INFO] [stderr] cargo:warning=./llama.cpp/llama.h:621:31: note: declared here [INFO] [stderr] cargo:warning= 621 | LLAMA_API DEPRECATED(void llama_sample_temperature( [INFO] [stderr] cargo:warning= | ^~~~~~~~~~~~~~~~~~~~~~~~ [INFO] [stderr] cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED' [INFO] [stderr] cargo:warning= 31 | # define DEPRECATED(func, hint) func __attribute__((deprecated(hint))) [INFO] [stderr] cargo:warning= | ^~~~ [INFO] [stderr] cargo:warning=./binding.cpp:456:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations] [INFO] [stderr] cargo:warning= 456 | llama_sample_temperature(ctx, &candidates_p, temp); [INFO] [stderr] cargo:warning= | ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~ [INFO] [stderr] cargo:warning=./llama.cpp/llama.h:621:31: note: declared here [INFO] [stderr] cargo:warning= 621 | LLAMA_API DEPRECATED(void llama_sample_temperature( [INFO] [stderr] cargo:warning= | ^~~~~~~~~~~~~~~~~~~~~~~~ [INFO] [stderr] cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED' [INFO] [stderr] cargo:warning= 31 | # define DEPRECATED(func, hint) func __attribute__((deprecated(hint))) [INFO] [stderr] cargo:warning= | ^~~~ [INFO] [stderr] cargo:warning=./binding.cpp:475:42: warning: cast from type 'const char*' to type 'char*' casts away qualifiers [-Wcast-qual] [INFO] [stderr] cargo:warning= 475 | if (!tokenCallback(state_pr, (char*)token_str.c_str())) [INFO] [stderr] cargo:warning= | ^~~~~~~~~~~~~~~~~~~~~~~~ [INFO] [stderr] cargo:warning=./binding.cpp: In function 'void* llama_allocate_params(const char*, int, int, int, int, float, float, float, int, bool, bool, int, int, const char**, int, float, float, float, float, int, float, float, bool, const char*, const char*, bool, bool, bool, const char*, const char*, bool)': [INFO] [stderr] cargo:warning=./binding.cpp:629:100: warning: unused parameter 'ignore_eos' [-Wunused-parameter] [INFO] [stderr] cargo:warning= 629 | float top_p, float temp, float repeat_penalty, int repeat_last_n, bool ignore_eos, bool memory_f16, int n_batch, int n_keep, const char **antiprompt, int antiprompt_count, [INFO] [stderr] cargo:warning= | ~~~~~^~~~~~~~~~ [INFO] [stderr] cargo:warning=./binding.cpp: In function 'void* load_model(const char*, int, int, bool, bool, bool, bool, bool, bool, int, int, const char*, const char*, bool)': [INFO] [stderr] cargo:warning=./binding.cpp:708:122: warning: unused parameter 'low_vram' [-Wunused-parameter] [INFO] [stderr] cargo:warning= 708 | void *load_model(const char *fname, int n_ctx, int n_seed, bool memory_f16, bool mlock, bool embeddings, bool mmap, bool low_vram, bool vocab_only, int n_gpu_layers, int n_batch, const char *maingpu, const char *tensorsplit, bool numa) [INFO] [stderr] cargo:warning= | ~~~~~^~~~~~~~ [INFO] [stderr] cargo:rerun-if-env-changed=AR_x86_64-unknown-linux-gnu [INFO] [stderr] AR_x86_64-unknown-linux-gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=AR_x86_64_unknown_linux_gnu [INFO] [stderr] AR_x86_64_unknown_linux_gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=HOST_AR [INFO] [stderr] HOST_AR = None [INFO] [stderr] cargo:rerun-if-env-changed=AR [INFO] [stderr] AR = None [INFO] [stderr] cargo:rerun-if-env-changed=ARFLAGS [INFO] [stderr] ARFLAGS = None [INFO] [stderr] cargo:rerun-if-env-changed=HOST_ARFLAGS [INFO] [stderr] HOST_ARFLAGS = None [INFO] [stderr] cargo:rerun-if-env-changed=ARFLAGS_x86_64_unknown_linux_gnu [INFO] [stderr] ARFLAGS_x86_64_unknown_linux_gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=ARFLAGS_x86_64-unknown-linux-gnu [INFO] [stderr] ARFLAGS_x86_64-unknown-linux-gnu = None [INFO] [stderr] cargo:warning=ar: /opt/rustwide/target/debug/build/llama_cpp_rs-f2e7c9ea871f58df/out/llama.cpp/ggml.o: No such file or directory [INFO] [stderr] [INFO] [stderr] --- stderr [INFO] [stderr] error: 'rustfmt' is not installed for the custom toolchain 'a4980311fb7bb9e7893708e6bd3fbbfb2819fd3d'. [INFO] [stderr] note: this is a custom toolchain, which cannot use `rustup component add` [INFO] [stderr] help: if you built this toolchain from source, and used `rustup toolchain link`, then you may be able to build the component with `x.py` [INFO] [stderr] Failed to run rustfmt: Internal rustfmt error (non-fatal, continuing) [INFO] [stderr] [INFO] [stderr] [INFO] [stderr] error occurred in cc-rs: command did not execute successfully (status code exit status: 1): ZERO_AR_DATE="1" "ar" "cq" "/opt/rustwide/target/debug/build/llama_cpp_rs-f2e7c9ea871f58df/out/libbinding.a" "/opt/rustwide/target/debug/build/llama_cpp_rs-f2e7c9ea871f58df/out/7db155ec2d396663-common.o" "/opt/rustwide/target/debug/build/llama_cpp_rs-f2e7c9ea871f58df/out/f326362f9e96224c-llama.o" "/opt/rustwide/target/debug/build/llama_cpp_rs-f2e7c9ea871f58df/out/c854047367d9b492-binding.o" "/opt/rustwide/target/debug/build/llama_cpp_rs-f2e7c9ea871f58df/out/llama.cpp/ggml.o" [INFO] [stderr] [INFO] [stderr] [INFO] running `Command { std: "docker" "inspect" "06c50779f00dbda1a3312e1a88d5c4a5bfaad4f8a668164588b4068290f336fe", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "06c50779f00dbda1a3312e1a88d5c4a5bfaad4f8a668164588b4068290f336fe", kill_on_drop: false }` [INFO] [stdout] 06c50779f00dbda1a3312e1a88d5c4a5bfaad4f8a668164588b4068290f336fe