[INFO] fetching crate llama_cpp_rs 0.3.0...
[INFO] testing llama_cpp_rs-0.3.0 against beta-2025-09-21 for beta-1.91-3
[INFO] extracting crate llama_cpp_rs 0.3.0 into /workspace/builds/worker-4-tc2/source
[INFO] started tweaking crates.io crate llama_cpp_rs 0.3.0
[INFO] finished tweaking crates.io crate llama_cpp_rs 0.3.0
[INFO] tweaked toml for crates.io crate llama_cpp_rs 0.3.0 written to /workspace/builds/worker-4-tc2/source/Cargo.toml
[INFO] validating manifest of crates.io crate llama_cpp_rs 0.3.0 on toolchain beta-2025-09-21
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+beta-2025-09-21" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }`
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+beta-2025-09-21" "generate-lockfile" "--manifest-path" "Cargo.toml", kill_on_drop: false }`
[INFO] [stderr]     Updating crates.io index
[INFO] [stderr]      Locking 47 packages to latest compatible versions
[INFO] [stderr]       Adding bindgen v0.66.1 (available: v0.72.1)
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+beta-2025-09-21" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-4-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-4-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:e90291280db7d1fac5b66fc6dad9f9662629e7365a55743daf9bdf73ebc4ea79" "/opt/rustwide/cargo-home/bin/cargo" "+beta-2025-09-21" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }`
[INFO] [stdout] 2d049400fbebc8ddf7578e7df5409f3d093133ec094ca569a40b54933c86c957
[INFO] running `Command { std: "docker" "start" "-a" "2d049400fbebc8ddf7578e7df5409f3d093133ec094ca569a40b54933c86c957", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "inspect" "2d049400fbebc8ddf7578e7df5409f3d093133ec094ca569a40b54933c86c957", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "2d049400fbebc8ddf7578e7df5409f3d093133ec094ca569a40b54933c86c957", kill_on_drop: false }`
[INFO] [stdout] 2d049400fbebc8ddf7578e7df5409f3d093133ec094ca569a40b54933c86c957
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-4-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-4-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=warn" "-e" "RUSTDOCFLAGS=--cap-lints=warn" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:e90291280db7d1fac5b66fc6dad9f9662629e7365a55743daf9bdf73ebc4ea79" "/opt/rustwide/cargo-home/bin/cargo" "+beta-2025-09-21" "build" "--frozen" "--message-format=json", kill_on_drop: false }`
[INFO] [stdout] 3951d4bfb557c5059618da39fa8d358b7db77fcec6830a0214af808527eb905d
[INFO] running `Command { std: "docker" "start" "-a" "3951d4bfb557c5059618da39fa8d358b7db77fcec6830a0214af808527eb905d", kill_on_drop: false }`
[INFO] [stderr]    Compiling prettyplease v0.2.37
[INFO] [stderr]    Compiling syn v2.0.106
[INFO] [stderr]    Compiling bindgen v0.66.1
[INFO] [stderr]    Compiling find-msvc-tools v0.1.2
[INFO] [stderr]    Compiling cc v1.2.39
[INFO] [stderr]    Compiling llama_cpp_rs v0.3.0 (/opt/rustwide/workdir)
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/ggml.c:19672:13: warning: 'ggml_opt_get_grad' defined but not used [-Wunused-function]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: 19672 | static void ggml_opt_get_grad(int np, struct ggml_tensor * const ps[], float * g) {
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |             ^~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/k_quants.c:205:14: warning: 'make_qkx1_quants' defined but not used [-Wunused-function]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   205 | static float make_qkx1_quants(int n, int nmax, const float * restrict x, uint8_t * restrict L, float * restrict the_min,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |              ^~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'int get_embeddings(void*, void*, float*)':
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:80:23: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    80 |         if (llama_eval(ctx, embd_inp.data(), embd_inp.size(), n_past))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |             ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                              ^~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'int eval(void*, void*, char*)':
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:138:22: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   138 |     return llama_eval(ctx, tokens.data(), n_prompt_tokens, n_past);
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |            ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                              ^~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'int llama_predict(void*, void*, char*, bool)':
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:282:19: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   282 |         llama_eval(ctx, tmp, 1, 0);
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |         ~~~~~~~~~~^~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                              ^~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:353:31: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   353 |                 if (llama_eval(ctx, &embd[i], n_eval, n_past))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                     ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                              ^~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:440:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   440 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:446:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   446 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:456:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   456 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:475:42: warning: cast from type 'const char*' to type 'char*' casts away qualifiers [-Wcast-qual]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   475 |             if (!tokenCallback(state_pr, (char*)token_str.c_str()))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                          ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'void* llama_allocate_params(const char*, int, int, int, int, float, float, float, int, bool, bool, int, int, const char**, int, float, float, float, float, int, float, float, bool, const char*, const char*, bool, bool, bool, const char*, const char*, bool)':
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:629:100: warning: unused parameter 'ignore_eos' [-Wunused-parameter]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   629 |                             float top_p, float temp, float repeat_penalty, int repeat_last_n, bool ignore_eos, bool memory_f16, int n_batch, int n_keep, const char **antiprompt, int antiprompt_count,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                                                                               ~~~~~^~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'void* load_model(const char*, int, int, bool, bool, bool, bool, bool, bool, int, int, const char*, const char*, bool)':
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:708:122: warning: unused parameter 'low_vram' [-Wunused-parameter]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   708 | void *load_model(const char *fname, int n_ctx, int n_seed, bool memory_f16, bool mlock, bool embeddings, bool mmap, bool low_vram, bool vocab_only, int n_gpu_layers, int n_batch, const char *maingpu, const char *tensorsplit, bool numa)
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                                                                                                     ~~~~~^~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ar: /opt/rustwide/target/debug/build/llama_cpp_rs-e5d17f432e9fa424/out/llama.cpp/ggml.o: No such file or directory
[INFO] [stderr] error: failed to run custom build command for `llama_cpp_rs v0.3.0 (/opt/rustwide/workdir)`
[INFO] [stderr] note: To improve backtraces for build dependencies, set the CARGO_PROFILE_DEV_BUILD_OVERRIDE_DEBUG=true environment variable to enable debug information generation.
[INFO] [stderr] 
[INFO] [stderr] Caused by:
[INFO] [stderr]   process didn't exit successfully: `/opt/rustwide/target/debug/build/llama_cpp_rs-84d61de3dedce643/build-script-build` (exit status: 1)
[INFO] [stderr]   --- stdout
[INFO] [stderr]   cargo:rerun-if-env-changed=TARGET
[INFO] [stderr]   cargo:rerun-if-env-changed=BINDGEN_EXTRA_CLANG_ARGS_x86_64-unknown-linux-gnu
[INFO] [stderr]   cargo:rerun-if-env-changed=BINDGEN_EXTRA_CLANG_ARGS_x86_64_unknown_linux_gnu
[INFO] [stderr]   cargo:rerun-if-env-changed=BINDGEN_EXTRA_CLANG_ARGS
[INFO] [stderr]   cargo:rerun-if-changed=/usr/lib/llvm-18/lib/clang/18/include/stdbool.h
[INFO] [stderr]   OUT_DIR = Some(/opt/rustwide/target/debug/build/llama_cpp_rs-e5d17f432e9fa424/out)
[INFO] [stderr]   OPT_LEVEL = Some(0)
[INFO] [stderr]   TARGET = Some(x86_64-unknown-linux-gnu)
[INFO] [stderr]   HOST = Some(x86_64-unknown-linux-gnu)
[INFO] [stderr]   cargo:rerun-if-env-changed=CC_x86_64-unknown-linux-gnu
[INFO] [stderr]   CC_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CC_x86_64_unknown_linux_gnu
[INFO] [stderr]   CC_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_CC
[INFO] [stderr]   HOST_CC = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CC
[INFO] [stderr]   CC = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CC_ENABLE_DEBUG_OUTPUT
[INFO] [stderr]   RUSTC_WRAPPER = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CRATE_CC_NO_DEFAULTS
[INFO] [stderr]   CRATE_CC_NO_DEFAULTS = None
[INFO] [stderr]   DEBUG = Some(true)
[INFO] [stderr]   CARGO_CFG_TARGET_FEATURE = Some(fxsr,sse,sse2)
[INFO] [stderr]   cargo:rerun-if-env-changed=CFLAGS
[INFO] [stderr]   CFLAGS = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_CFLAGS
[INFO] [stderr]   HOST_CFLAGS = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CFLAGS_x86_64_unknown_linux_gnu
[INFO] [stderr]   CFLAGS_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CFLAGS_x86_64-unknown-linux-gnu
[INFO] [stderr]   CFLAGS_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   CARGO_ENCODED_RUSTFLAGS = Some(--cap-lints=warn)
[INFO] [stderr]   cargo:warning=./llama.cpp/ggml.c:19672:13: warning: 'ggml_opt_get_grad' defined but not used [-Wunused-function]
[INFO] [stderr]   cargo:warning=19672 | static void ggml_opt_get_grad(int np, struct ggml_tensor * const ps[], float * g) {
[INFO] [stderr]   cargo:warning=      |             ^~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/k_quants.c:205:14: warning: 'make_qkx1_quants' defined but not used [-Wunused-function]
[INFO] [stderr]   cargo:warning=  205 | static float make_qkx1_quants(int n, int nmax, const float * restrict x, uint8_t * restrict L, float * restrict the_min,
[INFO] [stderr]   cargo:warning=      |              ^~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:rerun-if-env-changed=AR_x86_64-unknown-linux-gnu
[INFO] [stderr]   AR_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=AR_x86_64_unknown_linux_gnu
[INFO] [stderr]   AR_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_AR
[INFO] [stderr]   HOST_AR = None
[INFO] [stderr]   cargo:rerun-if-env-changed=AR
[INFO] [stderr]   AR = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS
[INFO] [stderr]   ARFLAGS = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_ARFLAGS
[INFO] [stderr]   HOST_ARFLAGS = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS_x86_64_unknown_linux_gnu
[INFO] [stderr]   ARFLAGS_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS_x86_64-unknown-linux-gnu
[INFO] [stderr]   ARFLAGS_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rustc-link-lib=static=ggml
[INFO] [stderr]   cargo:rustc-link-search=native=/opt/rustwide/target/debug/build/llama_cpp_rs-e5d17f432e9fa424/out
[INFO] [stderr]   OUT_DIR = Some(/opt/rustwide/target/debug/build/llama_cpp_rs-e5d17f432e9fa424/out)
[INFO] [stderr]   OPT_LEVEL = Some(0)
[INFO] [stderr]   TARGET = Some(x86_64-unknown-linux-gnu)
[INFO] [stderr]   HOST = Some(x86_64-unknown-linux-gnu)
[INFO] [stderr]   cargo:rerun-if-env-changed=CXX_x86_64-unknown-linux-gnu
[INFO] [stderr]   CXX_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CXX_x86_64_unknown_linux_gnu
[INFO] [stderr]   CXX_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_CXX
[INFO] [stderr]   HOST_CXX = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CXX
[INFO] [stderr]   CXX = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CC_ENABLE_DEBUG_OUTPUT
[INFO] [stderr]   RUSTC_WRAPPER = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CRATE_CC_NO_DEFAULTS
[INFO] [stderr]   CRATE_CC_NO_DEFAULTS = None
[INFO] [stderr]   DEBUG = Some(true)
[INFO] [stderr]   CARGO_CFG_TARGET_FEATURE = Some(fxsr,sse,sse2)
[INFO] [stderr]   cargo:rerun-if-env-changed=CXXFLAGS
[INFO] [stderr]   CXXFLAGS = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_CXXFLAGS
[INFO] [stderr]   HOST_CXXFLAGS = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CXXFLAGS_x86_64_unknown_linux_gnu
[INFO] [stderr]   CXXFLAGS_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CXXFLAGS_x86_64-unknown-linux-gnu
[INFO] [stderr]   CXXFLAGS_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   CARGO_ENCODED_RUSTFLAGS = Some(--cap-lints=warn)
[INFO] [stderr]   cargo:warning=./binding.cpp: In function 'int get_embeddings(void*, void*, float*)':
[INFO] [stderr]   cargo:warning=./binding.cpp:80:23: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=   80 |         if (llama_eval(ctx, embd_inp.data(), embd_inp.size(), n_past))
[INFO] [stderr]   cargo:warning=      |             ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr]   cargo:warning=  423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr]   cargo:warning=      |                              ^~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp: In function 'int eval(void*, void*, char*)':
[INFO] [stderr]   cargo:warning=./binding.cpp:138:22: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  138 |     return llama_eval(ctx, tokens.data(), n_prompt_tokens, n_past);
[INFO] [stderr]   cargo:warning=      |            ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr]   cargo:warning=  423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr]   cargo:warning=      |                              ^~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp: In function 'int llama_predict(void*, void*, char*, bool)':
[INFO] [stderr]   cargo:warning=./binding.cpp:282:19: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  282 |         llama_eval(ctx, tmp, 1, 0);
[INFO] [stderr]   cargo:warning=      |         ~~~~~~~~~~^~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr]   cargo:warning=  423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr]   cargo:warning=      |                              ^~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp:353:31: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  353 |                 if (llama_eval(ctx, &embd[i], n_eval, n_past))
[INFO] [stderr]   cargo:warning=      |                     ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr]   cargo:warning=  423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr]   cargo:warning=      |                              ^~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp:440:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  440 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr]   cargo:warning=      |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr]   cargo:warning=  621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr]   cargo:warning=      |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp:446:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  446 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr]   cargo:warning=      |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr]   cargo:warning=  621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr]   cargo:warning=      |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp:456:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  456 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr]   cargo:warning=      |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr]   cargo:warning=  621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr]   cargo:warning=      |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp:475:42: warning: cast from type 'const char*' to type 'char*' casts away qualifiers [-Wcast-qual]
[INFO] [stderr]   cargo:warning=  475 |             if (!tokenCallback(state_pr, (char*)token_str.c_str()))
[INFO] [stderr]   cargo:warning=      |                                          ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./binding.cpp: In function 'void* llama_allocate_params(const char*, int, int, int, int, float, float, float, int, bool, bool, int, int, const char**, int, float, float, float, float, int, float, float, bool, const char*, const char*, bool, bool, bool, const char*, const char*, bool)':
[INFO] [stderr]   cargo:warning=./binding.cpp:629:100: warning: unused parameter 'ignore_eos' [-Wunused-parameter]
[INFO] [stderr]   cargo:warning=  629 |                             float top_p, float temp, float repeat_penalty, int repeat_last_n, bool ignore_eos, bool memory_f16, int n_batch, int n_keep, const char **antiprompt, int antiprompt_count,
[INFO] [stderr]   cargo:warning=      |                                                                                               ~~~~~^~~~~~~~~~
[INFO] [stderr]   cargo:warning=./binding.cpp: In function 'void* load_model(const char*, int, int, bool, bool, bool, bool, bool, bool, int, int, const char*, const char*, bool)':
[INFO] [stderr]   cargo:warning=./binding.cpp:708:122: warning: unused parameter 'low_vram' [-Wunused-parameter]
[INFO] [stderr]   cargo:warning=  708 | void *load_model(const char *fname, int n_ctx, int n_seed, bool memory_f16, bool mlock, bool embeddings, bool mmap, bool low_vram, bool vocab_only, int n_gpu_layers, int n_batch, const char *maingpu, const char *tensorsplit, bool numa)
[INFO] [stderr]   cargo:warning=      |                                                                                                                     ~~~~~^~~~~~~~
[INFO] [stderr]   cargo:rerun-if-env-changed=AR_x86_64-unknown-linux-gnu
[INFO] [stderr]   AR_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=AR_x86_64_unknown_linux_gnu
[INFO] [stderr]   AR_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_AR
[INFO] [stderr]   HOST_AR = None
[INFO] [stderr]   cargo:rerun-if-env-changed=AR
[INFO] [stderr]   AR = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS
[INFO] [stderr]   ARFLAGS = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_ARFLAGS
[INFO] [stderr]   HOST_ARFLAGS = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS_x86_64_unknown_linux_gnu
[INFO] [stderr]   ARFLAGS_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS_x86_64-unknown-linux-gnu
[INFO] [stderr]   ARFLAGS_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:warning=ar: /opt/rustwide/target/debug/build/llama_cpp_rs-e5d17f432e9fa424/out/llama.cpp/ggml.o: No such file or directory
[INFO] [stderr] 
[INFO] [stderr]   --- stderr
[INFO] [stderr]   error: 'rustfmt' is not installed for the toolchain 'beta-2025-09-21-x86_64-unknown-linux-gnu'.
[INFO] [stderr]   To install, run `rustup component add --toolchain beta-2025-09-21-x86_64-unknown-linux-gnu rustfmt`
[INFO] [stderr]   Failed to run rustfmt: Internal rustfmt error (non-fatal, continuing)
[INFO] [stderr] 
[INFO] [stderr] 
[INFO] [stderr]   error occurred in cc-rs: command did not execute successfully (status code exit status: 1): ZERO_AR_DATE="1" "ar" "cq" "/opt/rustwide/target/debug/build/llama_cpp_rs-e5d17f432e9fa424/out/libbinding.a" "/opt/rustwide/target/debug/build/llama_cpp_rs-e5d17f432e9fa424/out/7db155ec2d396663-common.o" "/opt/rustwide/target/debug/build/llama_cpp_rs-e5d17f432e9fa424/out/f326362f9e96224c-llama.o" "/opt/rustwide/target/debug/build/llama_cpp_rs-e5d17f432e9fa424/out/c854047367d9b492-binding.o" "/opt/rustwide/target/debug/build/llama_cpp_rs-e5d17f432e9fa424/out/llama.cpp/ggml.o"
[INFO] [stderr] 
[INFO] [stderr] 
[INFO] running `Command { std: "docker" "inspect" "3951d4bfb557c5059618da39fa8d358b7db77fcec6830a0214af808527eb905d", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "3951d4bfb557c5059618da39fa8d358b7db77fcec6830a0214af808527eb905d", kill_on_drop: false }`
[INFO] [stdout] 3951d4bfb557c5059618da39fa8d358b7db77fcec6830a0214af808527eb905d
