[INFO] fetching crate llama_cpp_rs 0.3.0...
[INFO] checking llama_cpp_rs-0.3.0 against try#31561a63598a0f21fc74c024747ef1310fdd2eb2 for pr-133536
[INFO] extracting crate llama_cpp_rs 0.3.0 into /workspace/builds/worker-2-tc2/source
[INFO] validating manifest of crates.io crate llama_cpp_rs 0.3.0 on toolchain 31561a63598a0f21fc74c024747ef1310fdd2eb2
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+31561a63598a0f21fc74c024747ef1310fdd2eb2" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }`
[INFO] started tweaking crates.io crate llama_cpp_rs 0.3.0
[INFO] finished tweaking crates.io crate llama_cpp_rs 0.3.0
[INFO] tweaked toml for crates.io crate llama_cpp_rs 0.3.0 written to /workspace/builds/worker-2-tc2/source/Cargo.toml
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+31561a63598a0f21fc74c024747ef1310fdd2eb2" "generate-lockfile" "--manifest-path" "Cargo.toml", kill_on_drop: false }`
[INFO] [stderr]     Updating crates.io index
[INFO] [stderr]      Locking 45 packages to latest compatible versions
[INFO] [stderr]       Adding bindgen v0.66.1 (available: v0.70.1)
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+31561a63598a0f21fc74c024747ef1310fdd2eb2" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-2-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-2-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:4a844ea9eb2546a2d2c7022eacef16ef2e8229c7fbb2c7d4d55a9ceca922f72d" "/opt/rustwide/cargo-home/bin/cargo" "+31561a63598a0f21fc74c024747ef1310fdd2eb2" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }`
[INFO] [stdout] 9d4488ae268a845869db2758343df0f0c6bddb656120a44960e4f648e8b21a5f
[INFO] running `Command { std: "docker" "start" "-a" "9d4488ae268a845869db2758343df0f0c6bddb656120a44960e4f648e8b21a5f", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "inspect" "9d4488ae268a845869db2758343df0f0c6bddb656120a44960e4f648e8b21a5f", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "9d4488ae268a845869db2758343df0f0c6bddb656120a44960e4f648e8b21a5f", kill_on_drop: false }`
[INFO] [stdout] 9d4488ae268a845869db2758343df0f0c6bddb656120a44960e4f648e8b21a5f
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-2-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-2-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:4a844ea9eb2546a2d2c7022eacef16ef2e8229c7fbb2c7d4d55a9ceca922f72d" "/opt/rustwide/cargo-home/bin/cargo" "+31561a63598a0f21fc74c024747ef1310fdd2eb2" "check" "--frozen" "--all" "--all-targets" "--message-format=json", kill_on_drop: false }`
[INFO] [stdout] bb8983dee9aab273333ddfd7663886fc4512b1c9b14463c11a0ff847afda9da0
[INFO] running `Command { std: "docker" "start" "-a" "bb8983dee9aab273333ddfd7663886fc4512b1c9b14463c11a0ff847afda9da0", kill_on_drop: false }`
[INFO] [stderr]    Compiling glob v0.3.1
[INFO] [stderr]    Compiling prettyplease v0.2.25
[INFO] [stderr]    Compiling minimal-lexical v0.2.1
[INFO] [stderr]    Compiling linux-raw-sys v0.4.14
[INFO] [stderr]    Compiling rustix v0.38.41
[INFO] [stderr]    Compiling regex-syntax v0.8.5
[INFO] [stderr]    Compiling memchr v2.7.4
[INFO] [stderr]    Compiling libloading v0.8.6
[INFO] [stderr]    Compiling syn v2.0.90
[INFO] [stderr]    Compiling bindgen v0.66.1
[INFO] [stderr]    Compiling lazycell v1.3.0
[INFO] [stderr]    Compiling rustc-hash v1.1.0
[INFO] [stderr]    Compiling log v0.4.22
[INFO] [stderr]    Compiling clang-sys v1.8.1
[INFO] [stderr]    Compiling nom v7.1.3
[INFO] [stderr]    Compiling which v4.4.2
[INFO] [stderr]    Compiling cexpr v0.6.0
[INFO] [stderr]    Compiling regex-automata v0.4.9
[INFO] [stderr]    Compiling regex v1.11.1
[INFO] [stderr]    Compiling llama_cpp_rs v0.3.0 (/opt/rustwide/workdir)
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/ggml.c:19672:13: warning: 'ggml_opt_get_grad' defined but not used [-Wunused-function]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: 19672 | static void ggml_opt_get_grad(int np, struct ggml_tensor * const ps[], float * g) {
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |             ^~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/k_quants.c:205:14: warning: 'make_qkx1_quants' defined but not used [-Wunused-function]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   205 | static float make_qkx1_quants(int n, int nmax, const float * restrict x, uint8_t * restrict L, float * restrict the_min,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |              ^~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'int get_embeddings(void*, void*, float*)':
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:80:23: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    80 |         if (llama_eval(ctx, embd_inp.data(), embd_inp.size(), n_past))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |             ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                              ^~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'int eval(void*, void*, char*)':
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:138:22: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   138 |     return llama_eval(ctx, tokens.data(), n_prompt_tokens, n_past);
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |            ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                              ^~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'int llama_predict(void*, void*, char*, bool)':
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:282:19: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   282 |         llama_eval(ctx, tmp, 1, 0);
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |         ~~~~~~~~~~^~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                              ^~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:353:31: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   353 |                 if (llama_eval(ctx, &embd[i], n_eval, n_past))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                     ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                              ^~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:440:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   440 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:446:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   446 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:456:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   456 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:475:42: warning: cast from type 'const char*' to type 'char*' casts away qualifiers [-Wcast-qual]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   475 |             if (!tokenCallback(state_pr, (char*)token_str.c_str()))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                          ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'void* llama_allocate_params(const char*, int, int, int, int, float, float, float, int, bool, bool, int, int, const char**, int, float, float, float, float, int, float, float, bool, const char*, const char*, bool, bool, bool, const char*, const char*, bool)':
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:629:100: warning: unused parameter 'ignore_eos' [-Wunused-parameter]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   629 |                             float top_p, float temp, float repeat_penalty, int repeat_last_n, bool ignore_eos, bool memory_f16, int n_batch, int n_keep, const char **antiprompt, int antiprompt_count,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                                                                               ~~~~~^~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'void* load_model(const char*, int, int, bool, bool, bool, bool, bool, bool, int, int, const char*, const char*, bool)':
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:708:122: warning: unused parameter 'low_vram' [-Wunused-parameter]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   708 | void *load_model(const char *fname, int n_ctx, int n_seed, bool memory_f16, bool mlock, bool embeddings, bool mmap, bool low_vram, bool vocab_only, int n_gpu_layers, int n_batch, const char *maingpu, const char *tensorsplit, bool numa)
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                                                                                                     ~~~~~^~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ar: /opt/rustwide/target/debug/build/llama_cpp_rs-821d5fa15aa5e57f/out/llama.cpp/ggml.o: No such file or directory
[INFO] [stderr] error: failed to run custom build command for `llama_cpp_rs v0.3.0 (/opt/rustwide/workdir)`
[INFO] [stderr] note: To improve backtraces for build dependencies, set the CARGO_PROFILE_DEV_BUILD_OVERRIDE_DEBUG=true environment variable to enable debug information generation.
[INFO] [stderr] 
[INFO] [stderr] Caused by:
[INFO] [stderr]   process didn't exit successfully: `/opt/rustwide/target/debug/build/llama_cpp_rs-edf7ee443628e38f/build-script-build` (exit status: 1)
[INFO] [stderr]   --- stdout
[INFO] [stderr]   cargo:rerun-if-env-changed=TARGET
[INFO] [stderr]   cargo:rerun-if-env-changed=BINDGEN_EXTRA_CLANG_ARGS_x86_64-unknown-linux-gnu
[INFO] [stderr]   cargo:rerun-if-env-changed=BINDGEN_EXTRA_CLANG_ARGS_x86_64_unknown_linux_gnu
[INFO] [stderr]   cargo:rerun-if-env-changed=BINDGEN_EXTRA_CLANG_ARGS
[INFO] [stderr]   cargo:rerun-if-changed=/usr/lib/llvm-14/lib/clang/14.0.0/include/stdbool.h
[INFO] [stderr]   OUT_DIR = Some(/opt/rustwide/target/debug/build/llama_cpp_rs-821d5fa15aa5e57f/out)
[INFO] [stderr]   OPT_LEVEL = Some(0)
[INFO] [stderr]   TARGET = Some(x86_64-unknown-linux-gnu)
[INFO] [stderr]   HOST = Some(x86_64-unknown-linux-gnu)
[INFO] [stderr]   cargo:rerun-if-env-changed=CC_x86_64-unknown-linux-gnu
[INFO] [stderr]   CC_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CC_x86_64_unknown_linux_gnu
[INFO] [stderr]   CC_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_CC
[INFO] [stderr]   HOST_CC = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CC
[INFO] [stderr]   CC = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CC_ENABLE_DEBUG_OUTPUT
[INFO] [stderr]   RUSTC_WRAPPER = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CRATE_CC_NO_DEFAULTS
[INFO] [stderr]   CRATE_CC_NO_DEFAULTS = None
[INFO] [stderr]   DEBUG = Some(true)
[INFO] [stderr]   CARGO_CFG_TARGET_FEATURE = Some(fxsr,sse,sse2)
[INFO] [stderr]   cargo:rerun-if-env-changed=CFLAGS_x86_64-unknown-linux-gnu
[INFO] [stderr]   CFLAGS_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CFLAGS_x86_64_unknown_linux_gnu
[INFO] [stderr]   CFLAGS_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_CFLAGS
[INFO] [stderr]   HOST_CFLAGS = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CFLAGS
[INFO] [stderr]   CFLAGS = None
[INFO] [stderr]   CARGO_ENCODED_RUSTFLAGS = Some(--cap-lints=forbid)
[INFO] [stderr]   cargo:warning=./llama.cpp/ggml.c:19672:13: warning: 'ggml_opt_get_grad' defined but not used [-Wunused-function]
[INFO] [stderr]   cargo:warning=19672 | static void ggml_opt_get_grad(int np, struct ggml_tensor * const ps[], float * g) {
[INFO] [stderr]   cargo:warning=      |             ^~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/k_quants.c:205:14: warning: 'make_qkx1_quants' defined but not used [-Wunused-function]
[INFO] [stderr]   cargo:warning=  205 | static float make_qkx1_quants(int n, int nmax, const float * restrict x, uint8_t * restrict L, float * restrict the_min,
[INFO] [stderr]   cargo:warning=      |              ^~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:rerun-if-env-changed=AR_x86_64-unknown-linux-gnu
[INFO] [stderr]   AR_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=AR_x86_64_unknown_linux_gnu
[INFO] [stderr]   AR_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_AR
[INFO] [stderr]   HOST_AR = None
[INFO] [stderr]   cargo:rerun-if-env-changed=AR
[INFO] [stderr]   AR = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS_x86_64-unknown-linux-gnu
[INFO] [stderr]   ARFLAGS_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS_x86_64_unknown_linux_gnu
[INFO] [stderr]   ARFLAGS_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_ARFLAGS
[INFO] [stderr]   HOST_ARFLAGS = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS
[INFO] [stderr]   ARFLAGS = None
[INFO] [stderr]   cargo:rustc-link-lib=static=ggml
[INFO] [stderr]   cargo:rustc-link-search=native=/opt/rustwide/target/debug/build/llama_cpp_rs-821d5fa15aa5e57f/out
[INFO] [stderr]   OUT_DIR = Some(/opt/rustwide/target/debug/build/llama_cpp_rs-821d5fa15aa5e57f/out)
[INFO] [stderr]   OPT_LEVEL = Some(0)
[INFO] [stderr]   TARGET = Some(x86_64-unknown-linux-gnu)
[INFO] [stderr]   HOST = Some(x86_64-unknown-linux-gnu)
[INFO] [stderr]   cargo:rerun-if-env-changed=CXX_x86_64-unknown-linux-gnu
[INFO] [stderr]   CXX_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CXX_x86_64_unknown_linux_gnu
[INFO] [stderr]   CXX_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_CXX
[INFO] [stderr]   HOST_CXX = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CXX
[INFO] [stderr]   CXX = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CC_ENABLE_DEBUG_OUTPUT
[INFO] [stderr]   RUSTC_WRAPPER = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CRATE_CC_NO_DEFAULTS
[INFO] [stderr]   CRATE_CC_NO_DEFAULTS = None
[INFO] [stderr]   DEBUG = Some(true)
[INFO] [stderr]   CARGO_CFG_TARGET_FEATURE = Some(fxsr,sse,sse2)
[INFO] [stderr]   cargo:rerun-if-env-changed=CXXFLAGS_x86_64-unknown-linux-gnu
[INFO] [stderr]   CXXFLAGS_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CXXFLAGS_x86_64_unknown_linux_gnu
[INFO] [stderr]   CXXFLAGS_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_CXXFLAGS
[INFO] [stderr]   HOST_CXXFLAGS = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CXXFLAGS
[INFO] [stderr]   CXXFLAGS = None
[INFO] [stderr]   CARGO_ENCODED_RUSTFLAGS = Some(--cap-lints=forbid)
[INFO] [stderr]   cargo:warning=./binding.cpp: In function 'int get_embeddings(void*, void*, float*)':
[INFO] [stderr]   cargo:warning=./binding.cpp:80:23: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=   80 |         if (llama_eval(ctx, embd_inp.data(), embd_inp.size(), n_past))
[INFO] [stderr]   cargo:warning=      |             ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr]   cargo:warning=  423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr]   cargo:warning=      |                              ^~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp: In function 'int eval(void*, void*, char*)':
[INFO] [stderr]   cargo:warning=./binding.cpp:138:22: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  138 |     return llama_eval(ctx, tokens.data(), n_prompt_tokens, n_past);
[INFO] [stderr]   cargo:warning=      |            ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr]   cargo:warning=  423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr]   cargo:warning=      |                              ^~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp: In function 'int llama_predict(void*, void*, char*, bool)':
[INFO] [stderr]   cargo:warning=./binding.cpp:282:19: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  282 |         llama_eval(ctx, tmp, 1, 0);
[INFO] [stderr]   cargo:warning=      |         ~~~~~~~~~~^~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr]   cargo:warning=  423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr]   cargo:warning=      |                              ^~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp:353:31: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  353 |                 if (llama_eval(ctx, &embd[i], n_eval, n_past))
[INFO] [stderr]   cargo:warning=      |                     ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr]   cargo:warning=  423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr]   cargo:warning=      |                              ^~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp:440:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  440 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr]   cargo:warning=      |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr]   cargo:warning=  621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr]   cargo:warning=      |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp:446:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  446 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr]   cargo:warning=      |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr]   cargo:warning=  621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr]   cargo:warning=      |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp:456:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  456 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr]   cargo:warning=      |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr]   cargo:warning=  621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr]   cargo:warning=      |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp:475:42: warning: cast from type 'const char*' to type 'char*' casts away qualifiers [-Wcast-qual]
[INFO] [stderr]   cargo:warning=  475 |             if (!tokenCallback(state_pr, (char*)token_str.c_str()))
[INFO] [stderr]   cargo:warning=      |                                          ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./binding.cpp: In function 'void* llama_allocate_params(const char*, int, int, int, int, float, float, float, int, bool, bool, int, int, const char**, int, float, float, float, float, int, float, float, bool, const char*, const char*, bool, bool, bool, const char*, const char*, bool)':
[INFO] [stderr]   cargo:warning=./binding.cpp:629:100: warning: unused parameter 'ignore_eos' [-Wunused-parameter]
[INFO] [stderr]   cargo:warning=  629 |                             float top_p, float temp, float repeat_penalty, int repeat_last_n, bool ignore_eos, bool memory_f16, int n_batch, int n_keep, const char **antiprompt, int antiprompt_count,
[INFO] [stderr]   cargo:warning=      |                                                                                               ~~~~~^~~~~~~~~~
[INFO] [stderr]   cargo:warning=./binding.cpp: In function 'void* load_model(const char*, int, int, bool, bool, bool, bool, bool, bool, int, int, const char*, const char*, bool)':
[INFO] [stderr]   cargo:warning=./binding.cpp:708:122: warning: unused parameter 'low_vram' [-Wunused-parameter]
[INFO] [stderr]   cargo:warning=  708 | void *load_model(const char *fname, int n_ctx, int n_seed, bool memory_f16, bool mlock, bool embeddings, bool mmap, bool low_vram, bool vocab_only, int n_gpu_layers, int n_batch, const char *maingpu, const char *tensorsplit, bool numa)
[INFO] [stderr]   cargo:warning=      |                                                                                                                     ~~~~~^~~~~~~~
[INFO] [stderr]   cargo:rerun-if-env-changed=AR_x86_64-unknown-linux-gnu
[INFO] [stderr]   AR_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=AR_x86_64_unknown_linux_gnu
[INFO] [stderr]   AR_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_AR
[INFO] [stderr]   HOST_AR = None
[INFO] [stderr]   cargo:rerun-if-env-changed=AR
[INFO] [stderr]   AR = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS_x86_64-unknown-linux-gnu
[INFO] [stderr]   ARFLAGS_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS_x86_64_unknown_linux_gnu
[INFO] [stderr]   ARFLAGS_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_ARFLAGS
[INFO] [stderr]   HOST_ARFLAGS = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS
[INFO] [stderr]   ARFLAGS = None
[INFO] [stderr]   cargo:warning=ar: /opt/rustwide/target/debug/build/llama_cpp_rs-821d5fa15aa5e57f/out/llama.cpp/ggml.o: No such file or directory
[INFO] [stderr] 
[INFO] [stderr]   --- stderr
[INFO] [stderr]   error: 'rustfmt' is not installed for the custom toolchain '31561a63598a0f21fc74c024747ef1310fdd2eb2'.
[INFO] [stderr]   note: this is a custom toolchain, which cannot use `rustup component add`
[INFO] [stderr]   help: if you built this toolchain from source, and used `rustup toolchain link`, then you may be able to build the component with `x.py`
[INFO] [stderr]   Failed to run rustfmt: Internal rustfmt error (non-fatal, continuing)
[INFO] [stderr] 
[INFO] [stderr] 
[INFO] [stderr]   error occurred: Command ZERO_AR_DATE="1" "ar" "cq" "/opt/rustwide/target/debug/build/llama_cpp_rs-821d5fa15aa5e57f/out/libbinding.a" "/opt/rustwide/target/debug/build/llama_cpp_rs-821d5fa15aa5e57f/out/7db155ec2d396663-common.o" "/opt/rustwide/target/debug/build/llama_cpp_rs-821d5fa15aa5e57f/out/f326362f9e96224c-llama.o" "/opt/rustwide/target/debug/build/llama_cpp_rs-821d5fa15aa5e57f/out/c854047367d9b492-binding.o" "/opt/rustwide/target/debug/build/llama_cpp_rs-821d5fa15aa5e57f/out/llama.cpp/ggml.o" with args ar did not execute successfully (status code exit status: 1).
[INFO] [stderr] 
[INFO] [stderr] 
[INFO] running `Command { std: "docker" "inspect" "bb8983dee9aab273333ddfd7663886fc4512b1c9b14463c11a0ff847afda9da0", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "bb8983dee9aab273333ddfd7663886fc4512b1c9b14463c11a0ff847afda9da0", kill_on_drop: false }`
[INFO] [stdout] bb8983dee9aab273333ddfd7663886fc4512b1c9b14463c11a0ff847afda9da0
