[INFO] fetching crate llama_cpp_rs 0.3.0...
[INFO] checking llama_cpp_rs-0.3.0 against master#c2f2db79ca3024f68d22b45aa22b570775c2c4ad for pr-124157
[INFO] extracting crate llama_cpp_rs 0.3.0 into /workspace/builds/worker-1-tc1/source
[INFO] validating manifest of crates.io crate llama_cpp_rs 0.3.0 on toolchain c2f2db79ca3024f68d22b45aa22b570775c2c4ad
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+c2f2db79ca3024f68d22b45aa22b570775c2c4ad" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }`
[INFO] started tweaking crates.io crate llama_cpp_rs 0.3.0
[INFO] finished tweaking crates.io crate llama_cpp_rs 0.3.0
[INFO] tweaked toml for crates.io crate llama_cpp_rs 0.3.0 written to /workspace/builds/worker-1-tc1/source/Cargo.toml
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+c2f2db79ca3024f68d22b45aa22b570775c2c4ad" "generate-lockfile" "--manifest-path" "Cargo.toml", kill_on_drop: false }`
[INFO] [stderr]     Updating crates.io index
[INFO] [stderr]      Locking 45 packages to latest compatible versions
[INFO] [stderr]       Adding bindgen v0.66.1 (latest: v0.69.4)
[INFO] [stderr]       Adding linux-raw-sys v0.4.13 (latest: v0.6.4)
[INFO] [stderr]       Adding peeking_take_while v0.1.2 (latest: v1.0.0)
[INFO] [stderr]       Adding which v4.4.2 (latest: v6.0.1)
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+c2f2db79ca3024f68d22b45aa22b570775c2c4ad" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-1-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-1-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:99613afd962a8cfa530ec1899472a458bd015a1ab0af876cf7eb06f6006d81ea" "/opt/rustwide/cargo-home/bin/cargo" "+c2f2db79ca3024f68d22b45aa22b570775c2c4ad" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }`
[INFO] [stdout] ad255adf520bdb5bfaab34893fe3ecbeae5f47053a595f513ef9c5a2e1ed459f
[INFO] running `Command { std: "docker" "start" "-a" "ad255adf520bdb5bfaab34893fe3ecbeae5f47053a595f513ef9c5a2e1ed459f", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "inspect" "ad255adf520bdb5bfaab34893fe3ecbeae5f47053a595f513ef9c5a2e1ed459f", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "ad255adf520bdb5bfaab34893fe3ecbeae5f47053a595f513ef9c5a2e1ed459f", kill_on_drop: false }`
[INFO] [stdout] ad255adf520bdb5bfaab34893fe3ecbeae5f47053a595f513ef9c5a2e1ed459f
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-1-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-1-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:99613afd962a8cfa530ec1899472a458bd015a1ab0af876cf7eb06f6006d81ea" "/opt/rustwide/cargo-home/bin/cargo" "+c2f2db79ca3024f68d22b45aa22b570775c2c4ad" "check" "--frozen" "--all" "--all-targets" "--message-format=json", kill_on_drop: false }`
[INFO] [stdout] 93cdf59e40ad4410c26c328b02e3a0e9c70f87d575fc896b8102e4c7ca9c570c
[INFO] running `Command { std: "docker" "start" "-a" "93cdf59e40ad4410c26c328b02e3a0e9c70f87d575fc896b8102e4c7ca9c570c", kill_on_drop: false }`
[INFO] [stderr]    Compiling proc-macro2 v1.0.81
[INFO] [stderr]    Compiling libc v0.2.154
[INFO] [stderr]    Compiling prettyplease v0.2.19
[INFO] [stderr]    Compiling rustix v0.38.34
[INFO] [stderr]    Compiling linux-raw-sys v0.4.13
[INFO] [stderr]    Compiling clang-sys v1.7.0
[INFO] [stderr]    Compiling bitflags v2.5.0
[INFO] [stderr]    Compiling memchr v2.7.2
[INFO] [stderr]    Compiling regex-syntax v0.8.3
[INFO] [stderr]    Compiling libloading v0.8.3
[INFO] [stderr]    Compiling home v0.5.9
[INFO] [stderr]    Compiling either v1.11.0
[INFO] [stderr]    Compiling bindgen v0.66.1
[INFO] [stderr]    Compiling nom v7.1.3
[INFO] [stderr]    Compiling lazy_static v1.4.0
[INFO] [stderr]    Compiling quote v1.0.36
[INFO] [stderr]    Compiling peeking_take_while v0.1.2
[INFO] [stderr]    Compiling shlex v1.3.0
[INFO] [stderr]    Compiling log v0.4.21
[INFO] [stderr]    Compiling syn v2.0.60
[INFO] [stderr]    Compiling rustc-hash v1.1.0
[INFO] [stderr]    Compiling lazycell v1.3.0
[INFO] [stderr]    Compiling regex-automata v0.4.6
[INFO] [stderr]    Compiling cexpr v0.6.0
[INFO] [stderr]    Compiling which v4.4.2
[INFO] [stderr]    Compiling regex v1.10.4
[INFO] [stderr]    Compiling llama_cpp_rs v0.3.0 (/opt/rustwide/workdir)
[INFO] [stderr] The following warnings were emitted during compilation:
[INFO] [stderr] 
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/ggml.c:19672:13: warning: 'ggml_opt_get_grad' defined but not used [-Wunused-function]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: 19672 | static void ggml_opt_get_grad(int np, struct ggml_tensor * const ps[], float * g) {
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |             ^~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/k_quants.c:205:14: warning: 'make_qkx1_quants' defined but not used [-Wunused-function]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   205 | static float make_qkx1_quants(int n, int nmax, const float * restrict x, uint8_t * restrict L, float * restrict the_min,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |              ^~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'int get_embeddings(void*, void*, float*)':
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:80:23: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    80 |         if (llama_eval(ctx, embd_inp.data(), embd_inp.size(), n_past))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |             ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                              ^~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'int eval(void*, void*, char*)':
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:138:22: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   138 |     return llama_eval(ctx, tokens.data(), n_prompt_tokens, n_past);
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |            ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                              ^~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'int llama_predict(void*, void*, char*, bool)':
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:282:19: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   282 |         llama_eval(ctx, tmp, 1, 0);
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |         ~~~~~~~~~~^~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                              ^~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:353:31: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   353 |                 if (llama_eval(ctx, &embd[i], n_eval, n_past))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                     ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                              ^~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:440:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   440 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:446:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   446 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:456:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   456 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:475:42: warning: cast from type 'const char*' to type 'char*' casts away qualifiers [-Wcast-qual]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   475 |             if (!tokenCallback(state_pr, (char*)token_str.c_str()))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                          ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'void* llama_allocate_params(const char*, int, int, int, int, float, float, float, int, bool, bool, int, int, const char**, int, float, float, float, float, int, float, float, bool, const char*, const char*, bool, bool, bool, const char*, const char*, bool)':
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:629:100: warning: unused parameter 'ignore_eos' [-Wunused-parameter]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   629 |                             float top_p, float temp, float repeat_penalty, int repeat_last_n, bool ignore_eos, bool memory_f16, int n_batch, int n_keep, const char **antiprompt, int antiprompt_count,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                                                                               ~~~~~^~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'void* load_model(const char*, int, int, bool, bool, bool, bool, bool, bool, int, int, const char*, const char*, bool)':
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:708:122: warning: unused parameter 'low_vram' [-Wunused-parameter]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   708 | void *load_model(const char *fname, int n_ctx, int n_seed, bool memory_f16, bool mlock, bool embeddings, bool mmap, bool low_vram, bool vocab_only, int n_gpu_layers, int n_batch, const char *maingpu, const char *tensorsplit, bool numa)
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                                                                                                     ~~~~~^~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ar: /opt/rustwide/target/debug/build/llama_cpp_rs-f2214abd2a37516b/out/llama.cpp/ggml.o: No such file or directory
[INFO] [stderr] 
[INFO] [stderr] error: failed to run custom build command for `llama_cpp_rs v0.3.0 (/opt/rustwide/workdir)`
[INFO] [stderr] note: To improve backtraces for build dependencies, set the CARGO_PROFILE_DEV_BUILD_OVERRIDE_DEBUG=true environment variable to enable debug information generation.
[INFO] [stderr] 
[INFO] [stderr] Caused by:
[INFO] [stderr]   process didn't exit successfully: `/opt/rustwide/target/debug/build/llama_cpp_rs-b3f679a9726fdee3/build-script-build` (exit status: 1)
[INFO] [stderr]   --- stdout
[INFO] [stderr]   cargo:rerun-if-env-changed=TARGET
[INFO] [stderr]   cargo:rerun-if-env-changed=BINDGEN_EXTRA_CLANG_ARGS_x86_64-unknown-linux-gnu
[INFO] [stderr]   cargo:rerun-if-env-changed=BINDGEN_EXTRA_CLANG_ARGS_x86_64_unknown_linux_gnu
[INFO] [stderr]   cargo:rerun-if-env-changed=BINDGEN_EXTRA_CLANG_ARGS
[INFO] [stderr]   cargo:rerun-if-changed=/usr/lib/llvm-14/lib/clang/14.0.0/include/stdbool.h
[INFO] [stderr]   TARGET = Some("x86_64-unknown-linux-gnu")
[INFO] [stderr]   OPT_LEVEL = Some("0")
[INFO] [stderr]   HOST = Some("x86_64-unknown-linux-gnu")
[INFO] [stderr]   cargo:rerun-if-env-changed=CC_x86_64-unknown-linux-gnu
[INFO] [stderr]   CC_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CC_x86_64_unknown_linux_gnu
[INFO] [stderr]   CC_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_CC
[INFO] [stderr]   HOST_CC = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CC
[INFO] [stderr]   CC = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CC_ENABLE_DEBUG_OUTPUT
[INFO] [stderr]   cargo:rerun-if-env-changed=CRATE_CC_NO_DEFAULTS
[INFO] [stderr]   CRATE_CC_NO_DEFAULTS = None
[INFO] [stderr]   DEBUG = Some("true")
[INFO] [stderr]   CARGO_CFG_TARGET_FEATURE = Some("fxsr,sse,sse2")
[INFO] [stderr]   cargo:rerun-if-env-changed=CFLAGS_x86_64-unknown-linux-gnu
[INFO] [stderr]   CFLAGS_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CFLAGS_x86_64_unknown_linux_gnu
[INFO] [stderr]   CFLAGS_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_CFLAGS
[INFO] [stderr]   HOST_CFLAGS = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CFLAGS
[INFO] [stderr]   CFLAGS = None
[INFO] [stderr]   cargo:warning=./llama.cpp/ggml.c:19672:13: warning: 'ggml_opt_get_grad' defined but not used [-Wunused-function]
[INFO] [stderr]   cargo:warning=19672 | static void ggml_opt_get_grad(int np, struct ggml_tensor * const ps[], float * g) {
[INFO] [stderr]   cargo:warning=      |             ^~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/k_quants.c:205:14: warning: 'make_qkx1_quants' defined but not used [-Wunused-function]
[INFO] [stderr]   cargo:warning=  205 | static float make_qkx1_quants(int n, int nmax, const float * restrict x, uint8_t * restrict L, float * restrict the_min,
[INFO] [stderr]   cargo:warning=      |              ^~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:rerun-if-env-changed=AR_x86_64-unknown-linux-gnu
[INFO] [stderr]   AR_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=AR_x86_64_unknown_linux_gnu
[INFO] [stderr]   AR_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_AR
[INFO] [stderr]   HOST_AR = None
[INFO] [stderr]   cargo:rerun-if-env-changed=AR
[INFO] [stderr]   AR = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS_x86_64-unknown-linux-gnu
[INFO] [stderr]   ARFLAGS_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS_x86_64_unknown_linux_gnu
[INFO] [stderr]   ARFLAGS_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_ARFLAGS
[INFO] [stderr]   HOST_ARFLAGS = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS
[INFO] [stderr]   ARFLAGS = None
[INFO] [stderr]   cargo:rustc-link-lib=static=ggml
[INFO] [stderr]   cargo:rustc-link-search=native=/opt/rustwide/target/debug/build/llama_cpp_rs-f2214abd2a37516b/out
[INFO] [stderr]   TARGET = Some("x86_64-unknown-linux-gnu")
[INFO] [stderr]   OPT_LEVEL = Some("0")
[INFO] [stderr]   HOST = Some("x86_64-unknown-linux-gnu")
[INFO] [stderr]   cargo:rerun-if-env-changed=CXX_x86_64-unknown-linux-gnu
[INFO] [stderr]   CXX_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CXX_x86_64_unknown_linux_gnu
[INFO] [stderr]   CXX_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_CXX
[INFO] [stderr]   HOST_CXX = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CXX
[INFO] [stderr]   CXX = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CC_ENABLE_DEBUG_OUTPUT
[INFO] [stderr]   cargo:rerun-if-env-changed=CRATE_CC_NO_DEFAULTS
[INFO] [stderr]   CRATE_CC_NO_DEFAULTS = None
[INFO] [stderr]   DEBUG = Some("true")
[INFO] [stderr]   CARGO_CFG_TARGET_FEATURE = Some("fxsr,sse,sse2")
[INFO] [stderr]   cargo:rerun-if-env-changed=CXXFLAGS_x86_64-unknown-linux-gnu
[INFO] [stderr]   CXXFLAGS_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CXXFLAGS_x86_64_unknown_linux_gnu
[INFO] [stderr]   CXXFLAGS_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_CXXFLAGS
[INFO] [stderr]   HOST_CXXFLAGS = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CXXFLAGS
[INFO] [stderr]   CXXFLAGS = None
[INFO] [stderr]   cargo:warning=./binding.cpp: In function 'int get_embeddings(void*, void*, float*)':
[INFO] [stderr]   cargo:warning=./binding.cpp:80:23: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=   80 |         if (llama_eval(ctx, embd_inp.data(), embd_inp.size(), n_past))
[INFO] [stderr]   cargo:warning=      |             ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr]   cargo:warning=  423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr]   cargo:warning=      |                              ^~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp: In function 'int eval(void*, void*, char*)':
[INFO] [stderr]   cargo:warning=./binding.cpp:138:22: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  138 |     return llama_eval(ctx, tokens.data(), n_prompt_tokens, n_past);
[INFO] [stderr]   cargo:warning=      |            ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr]   cargo:warning=  423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr]   cargo:warning=      |                              ^~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp: In function 'int llama_predict(void*, void*, char*, bool)':
[INFO] [stderr]   cargo:warning=./binding.cpp:282:19: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  282 |         llama_eval(ctx, tmp, 1, 0);
[INFO] [stderr]   cargo:warning=      |         ~~~~~~~~~~^~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr]   cargo:warning=  423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr]   cargo:warning=      |                              ^~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp:353:31: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  353 |                 if (llama_eval(ctx, &embd[i], n_eval, n_past))
[INFO] [stderr]   cargo:warning=      |                     ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr]   cargo:warning=  423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr]   cargo:warning=      |                              ^~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp:440:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  440 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr]   cargo:warning=      |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr]   cargo:warning=  621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr]   cargo:warning=      |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp:446:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  446 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr]   cargo:warning=      |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr]   cargo:warning=  621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr]   cargo:warning=      |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp:456:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  456 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr]   cargo:warning=      |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr]   cargo:warning=  621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr]   cargo:warning=      |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp:475:42: warning: cast from type 'const char*' to type 'char*' casts away qualifiers [-Wcast-qual]
[INFO] [stderr]   cargo:warning=  475 |             if (!tokenCallback(state_pr, (char*)token_str.c_str()))
[INFO] [stderr]   cargo:warning=      |                                          ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./binding.cpp: In function 'void* llama_allocate_params(const char*, int, int, int, int, float, float, float, int, bool, bool, int, int, const char**, int, float, float, float, float, int, float, float, bool, const char*, const char*, bool, bool, bool, const char*, const char*, bool)':
[INFO] [stderr]   cargo:warning=./binding.cpp:629:100: warning: unused parameter 'ignore_eos' [-Wunused-parameter]
[INFO] [stderr]   cargo:warning=  629 |                             float top_p, float temp, float repeat_penalty, int repeat_last_n, bool ignore_eos, bool memory_f16, int n_batch, int n_keep, const char **antiprompt, int antiprompt_count,
[INFO] [stderr]   cargo:warning=      |                                                                                               ~~~~~^~~~~~~~~~
[INFO] [stderr]   cargo:warning=./binding.cpp: In function 'void* load_model(const char*, int, int, bool, bool, bool, bool, bool, bool, int, int, const char*, const char*, bool)':
[INFO] [stderr]   cargo:warning=./binding.cpp:708:122: warning: unused parameter 'low_vram' [-Wunused-parameter]
[INFO] [stderr]   cargo:warning=  708 | void *load_model(const char *fname, int n_ctx, int n_seed, bool memory_f16, bool mlock, bool embeddings, bool mmap, bool low_vram, bool vocab_only, int n_gpu_layers, int n_batch, const char *maingpu, const char *tensorsplit, bool numa)
[INFO] [stderr]   cargo:warning=      |                                                                                                                     ~~~~~^~~~~~~~
[INFO] [stderr]   cargo:rerun-if-env-changed=AR_x86_64-unknown-linux-gnu
[INFO] [stderr]   AR_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=AR_x86_64_unknown_linux_gnu
[INFO] [stderr]   AR_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_AR
[INFO] [stderr]   HOST_AR = None
[INFO] [stderr]   cargo:rerun-if-env-changed=AR
[INFO] [stderr]   AR = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS_x86_64-unknown-linux-gnu
[INFO] [stderr]   ARFLAGS_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS_x86_64_unknown_linux_gnu
[INFO] [stderr]   ARFLAGS_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_ARFLAGS
[INFO] [stderr]   HOST_ARFLAGS = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS
[INFO] [stderr]   ARFLAGS = None
[INFO] [stderr]   cargo:warning=ar: /opt/rustwide/target/debug/build/llama_cpp_rs-f2214abd2a37516b/out/llama.cpp/ggml.o: No such file or directory
[INFO] [stderr] 
[INFO] [stderr]   --- stderr
[INFO] [stderr]   error: 'rustfmt' is not installed for the custom toolchain 'c2f2db79ca3024f68d22b45aa22b570775c2c4ad'.
[INFO] [stderr]   note: this is a custom toolchain, which cannot use `rustup component add`
[INFO] [stderr]   help: if you built this toolchain from source, and used `rustup toolchain link`, then you may be able to build the component with `x.py`
[INFO] [stderr]   Failed to run rustfmt: Internal rustfmt error (non-fatal, continuing)
[INFO] [stderr] 
[INFO] [stderr] 
[INFO] [stderr]   error occurred: Command ZERO_AR_DATE="1" "ar" "cq" "/opt/rustwide/target/debug/build/llama_cpp_rs-f2214abd2a37516b/out/libbinding.a" "/opt/rustwide/target/debug/build/llama_cpp_rs-f2214abd2a37516b/out/073db387043af495-common.o" "/opt/rustwide/target/debug/build/llama_cpp_rs-f2214abd2a37516b/out/30b5508d68fcb5a8-llama.o" "/opt/rustwide/target/debug/build/llama_cpp_rs-f2214abd2a37516b/out/8f1a5a601f45df90-binding.o" "/opt/rustwide/target/debug/build/llama_cpp_rs-f2214abd2a37516b/out/llama.cpp/ggml.o" with args ar did not execute successfully (status code exit status: 1).
[INFO] [stderr] 
[INFO] [stderr] 
[INFO] running `Command { std: "docker" "inspect" "93cdf59e40ad4410c26c328b02e3a0e9c70f87d575fc896b8102e4c7ca9c570c", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "93cdf59e40ad4410c26c328b02e3a0e9c70f87d575fc896b8102e4c7ca9c570c", kill_on_drop: false }`
[INFO] [stdout] 93cdf59e40ad4410c26c328b02e3a0e9c70f87d575fc896b8102e4c7ca9c570c
