[INFO] fetching crate llama_cpp_rs 0.3.0...
[INFO] checking llama_cpp_rs-0.3.0 against master#30f74ff0dc4d66debc8b50724c446f817e5f75f4 for pr-121848-1
[INFO] extracting crate llama_cpp_rs 0.3.0 into /workspace/builds/worker-4-tc1/source
[INFO] validating manifest of crates.io crate llama_cpp_rs 0.3.0 on toolchain 30f74ff0dc4d66debc8b50724c446f817e5f75f4
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+30f74ff0dc4d66debc8b50724c446f817e5f75f4" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }`
[INFO] started tweaking crates.io crate llama_cpp_rs 0.3.0
[INFO] finished tweaking crates.io crate llama_cpp_rs 0.3.0
[INFO] tweaked toml for crates.io crate llama_cpp_rs 0.3.0 written to /workspace/builds/worker-4-tc1/source/Cargo.toml
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+30f74ff0dc4d66debc8b50724c446f817e5f75f4" "generate-lockfile" "--manifest-path" "Cargo.toml", kill_on_drop: false }`
[INFO] [stderr]     Updating crates.io index
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+30f74ff0dc4d66debc8b50724c446f817e5f75f4" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-4-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-4-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:2788e3201cd34a07e3172128adcd8b3090168a8e3bcc40d7c032b9dda1df7d1c" "/opt/rustwide/cargo-home/bin/cargo" "+30f74ff0dc4d66debc8b50724c446f817e5f75f4" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }`
[INFO] [stdout] 72e05b1342e2bd529e06978ff982292bc6cb78be0918c855dbfca99fd2d13bd0
[INFO] running `Command { std: "docker" "start" "-a" "72e05b1342e2bd529e06978ff982292bc6cb78be0918c855dbfca99fd2d13bd0", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "inspect" "72e05b1342e2bd529e06978ff982292bc6cb78be0918c855dbfca99fd2d13bd0", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "72e05b1342e2bd529e06978ff982292bc6cb78be0918c855dbfca99fd2d13bd0", kill_on_drop: false }`
[INFO] [stdout] 72e05b1342e2bd529e06978ff982292bc6cb78be0918c855dbfca99fd2d13bd0
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-4-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-4-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:2788e3201cd34a07e3172128adcd8b3090168a8e3bcc40d7c032b9dda1df7d1c" "/opt/rustwide/cargo-home/bin/cargo" "+30f74ff0dc4d66debc8b50724c446f817e5f75f4" "check" "--frozen" "--all" "--all-targets" "--message-format=json", kill_on_drop: false }`
[INFO] [stdout] 668684f68d7d56c1759a49a4780e1e55279a2bafa338c71e0452202f47e0c601
[INFO] running `Command { std: "docker" "start" "-a" "668684f68d7d56c1759a49a4780e1e55279a2bafa338c71e0452202f47e0c601", kill_on_drop: false }`
[INFO] [stderr]    Compiling glob v0.3.1
[INFO] [stderr]    Compiling rustix v0.38.32
[INFO] [stderr]    Compiling prettyplease v0.2.17
[INFO] [stderr]    Compiling bitflags v2.5.0
[INFO] [stderr]    Compiling proc-macro2 v1.0.79
[INFO] [stderr]    Compiling libc v0.2.153
[INFO] [stderr]    Compiling linux-raw-sys v0.4.13
[INFO] [stderr]    Compiling minimal-lexical v0.2.1
[INFO] [stderr]    Compiling regex-syntax v0.8.3
[INFO] [stderr]    Compiling nom v7.1.3
[INFO] [stderr]    Compiling libloading v0.8.3
[INFO] [stderr]    Compiling clang-sys v1.7.0
[INFO] [stderr]    Compiling quote v1.0.35
[INFO] [stderr]    Compiling either v1.10.0
[INFO] [stderr]    Compiling syn v2.0.57
[INFO] [stderr]    Compiling bindgen v0.66.1
[INFO] [stderr]    Compiling home v0.5.9
[INFO] [stderr]    Compiling shlex v1.3.0
[INFO] [stderr]    Compiling peeking_take_while v0.1.2
[INFO] [stderr]    Compiling lazy_static v1.4.0
[INFO] [stderr]    Compiling rustc-hash v1.1.0
[INFO] [stderr]    Compiling lazycell v1.3.0
[INFO] [stderr]    Compiling regex-automata v0.4.6
[INFO] [stderr]    Compiling which v4.4.2
[INFO] [stderr]    Compiling cexpr v0.6.0
[INFO] [stderr]    Compiling regex v1.10.4
[INFO] [stderr]    Compiling llama_cpp_rs v0.3.0 (/opt/rustwide/workdir)
[INFO] [stderr] The following warnings were emitted during compilation:
[INFO] [stderr] 
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: Compiler version doesn't include clang or GCC: "cc" "--version"
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/ggml.c:19672:13: warning: 'ggml_opt_get_grad' defined but not used [-Wunused-function]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: 19672 | static void ggml_opt_get_grad(int np, struct ggml_tensor * const ps[], float * g) {
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |             ^~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/k_quants.c:205:14: warning: 'make_qkx1_quants' defined but not used [-Wunused-function]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   205 | static float make_qkx1_quants(int n, int nmax, const float * restrict x, uint8_t * restrict L, float * restrict the_min,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |              ^~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: Compiler version doesn't include clang or GCC: "c++" "--version"
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'int get_embeddings(void*, void*, float*)':
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:80:23: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    80 |         if (llama_eval(ctx, embd_inp.data(), embd_inp.size(), n_past))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |             ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                              ^~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'int eval(void*, void*, char*)':
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:138:22: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   138 |     return llama_eval(ctx, tokens.data(), n_prompt_tokens, n_past);
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |            ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                              ^~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'int llama_predict(void*, void*, char*, bool)':
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:282:19: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   282 |         llama_eval(ctx, tmp, 1, 0);
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |         ~~~~~~~~~~^~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                              ^~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:353:31: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   353 |                 if (llama_eval(ctx, &embd[i], n_eval, n_past))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                     ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                              ^~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:440:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   440 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:446:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   446 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:456:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   456 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:475:42: warning: cast from type 'const char*' to type 'char*' casts away qualifiers [-Wcast-qual]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   475 |             if (!tokenCallback(state_pr, (char*)token_str.c_str()))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                          ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'void* llama_allocate_params(const char*, int, int, int, int, float, float, float, int, bool, bool, int, int, const char**, int, float, float, float, float, int, float, float, bool, const char*, const char*, bool, bool, bool, const char*, const char*, bool)':
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:629:100: warning: unused parameter 'ignore_eos' [-Wunused-parameter]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   629 |                             float top_p, float temp, float repeat_penalty, int repeat_last_n, bool ignore_eos, bool memory_f16, int n_batch, int n_keep, const char **antiprompt, int antiprompt_count,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                                                                               ~~~~~^~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'void* load_model(const char*, int, int, bool, bool, bool, bool, bool, bool, int, int, const char*, const char*, bool)':
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:708:122: warning: unused parameter 'low_vram' [-Wunused-parameter]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   708 | void *load_model(const char *fname, int n_ctx, int n_seed, bool memory_f16, bool mlock, bool embeddings, bool mmap, bool low_vram, bool vocab_only, int n_gpu_layers, int n_batch, const char *maingpu, const char *tensorsplit, bool numa)
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                                                                                                     ~~~~~^~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ar: /opt/rustwide/target/debug/build/llama_cpp_rs-a791689fede158f0/out/llama.cpp/ggml.o: No such file or directory
[INFO] [stderr] 
[INFO] [stderr] error: failed to run custom build command for `llama_cpp_rs v0.3.0 (/opt/rustwide/workdir)`
[INFO] [stderr] note: To improve backtraces for build dependencies, set the CARGO_PROFILE_DEV_BUILD_OVERRIDE_DEBUG=true environment variable to enable debug information generation.
[INFO] [stderr] 
[INFO] [stderr] Caused by:
[INFO] [stderr]   process didn't exit successfully: `/opt/rustwide/target/debug/build/llama_cpp_rs-3d7548e85adc98be/build-script-build` (exit status: 1)
[INFO] [stderr]   --- stdout
[INFO] [stderr]   cargo:rerun-if-env-changed=TARGET
[INFO] [stderr]   cargo:rerun-if-env-changed=BINDGEN_EXTRA_CLANG_ARGS_x86_64-unknown-linux-gnu
[INFO] [stderr]   cargo:rerun-if-env-changed=BINDGEN_EXTRA_CLANG_ARGS_x86_64_unknown_linux_gnu
[INFO] [stderr]   cargo:rerun-if-env-changed=BINDGEN_EXTRA_CLANG_ARGS
[INFO] [stderr]   cargo:rerun-if-changed=/usr/lib/llvm-14/lib/clang/14.0.0/include/stdbool.h
[INFO] [stderr]   TARGET = Some("x86_64-unknown-linux-gnu")
[INFO] [stderr]   OPT_LEVEL = Some("0")
[INFO] [stderr]   HOST = Some("x86_64-unknown-linux-gnu")
[INFO] [stderr]   cargo:rerun-if-env-changed=CC_x86_64-unknown-linux-gnu
[INFO] [stderr]   CC_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CC_x86_64_unknown_linux_gnu
[INFO] [stderr]   CC_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_CC
[INFO] [stderr]   HOST_CC = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CC
[INFO] [stderr]   CC = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CC_ENABLE_DEBUG_OUTPUT
[INFO] [stderr]   cargo:warning=Compiler version doesn't include clang or GCC: "cc" "--version"
[INFO] [stderr]   cargo:rerun-if-env-changed=CRATE_CC_NO_DEFAULTS
[INFO] [stderr]   CRATE_CC_NO_DEFAULTS = None
[INFO] [stderr]   DEBUG = Some("true")
[INFO] [stderr]   CARGO_CFG_TARGET_FEATURE = Some("fxsr,sse,sse2")
[INFO] [stderr]   cargo:rerun-if-env-changed=CFLAGS_x86_64-unknown-linux-gnu
[INFO] [stderr]   CFLAGS_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CFLAGS_x86_64_unknown_linux_gnu
[INFO] [stderr]   CFLAGS_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_CFLAGS
[INFO] [stderr]   HOST_CFLAGS = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CFLAGS
[INFO] [stderr]   CFLAGS = None
[INFO] [stderr]   cargo:warning=./llama.cpp/ggml.c:19672:13: warning: 'ggml_opt_get_grad' defined but not used [-Wunused-function]
[INFO] [stderr]   cargo:warning=19672 | static void ggml_opt_get_grad(int np, struct ggml_tensor * const ps[], float * g) {
[INFO] [stderr]   cargo:warning=      |             ^~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/k_quants.c:205:14: warning: 'make_qkx1_quants' defined but not used [-Wunused-function]
[INFO] [stderr]   cargo:warning=  205 | static float make_qkx1_quants(int n, int nmax, const float * restrict x, uint8_t * restrict L, float * restrict the_min,
[INFO] [stderr]   cargo:warning=      |              ^~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:rerun-if-env-changed=AR_x86_64-unknown-linux-gnu
[INFO] [stderr]   AR_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=AR_x86_64_unknown_linux_gnu
[INFO] [stderr]   AR_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_AR
[INFO] [stderr]   HOST_AR = None
[INFO] [stderr]   cargo:rerun-if-env-changed=AR
[INFO] [stderr]   AR = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS_x86_64-unknown-linux-gnu
[INFO] [stderr]   ARFLAGS_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS_x86_64_unknown_linux_gnu
[INFO] [stderr]   ARFLAGS_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_ARFLAGS
[INFO] [stderr]   HOST_ARFLAGS = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS
[INFO] [stderr]   ARFLAGS = None
[INFO] [stderr]   cargo:rustc-link-lib=static=ggml
[INFO] [stderr]   cargo:rustc-link-search=native=/opt/rustwide/target/debug/build/llama_cpp_rs-a791689fede158f0/out
[INFO] [stderr]   TARGET = Some("x86_64-unknown-linux-gnu")
[INFO] [stderr]   OPT_LEVEL = Some("0")
[INFO] [stderr]   HOST = Some("x86_64-unknown-linux-gnu")
[INFO] [stderr]   cargo:rerun-if-env-changed=CXX_x86_64-unknown-linux-gnu
[INFO] [stderr]   CXX_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CXX_x86_64_unknown_linux_gnu
[INFO] [stderr]   CXX_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_CXX
[INFO] [stderr]   HOST_CXX = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CXX
[INFO] [stderr]   CXX = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CC_ENABLE_DEBUG_OUTPUT
[INFO] [stderr]   cargo:warning=Compiler version doesn't include clang or GCC: "c++" "--version"
[INFO] [stderr]   cargo:rerun-if-env-changed=CRATE_CC_NO_DEFAULTS
[INFO] [stderr]   CRATE_CC_NO_DEFAULTS = None
[INFO] [stderr]   DEBUG = Some("true")
[INFO] [stderr]   CARGO_CFG_TARGET_FEATURE = Some("fxsr,sse,sse2")
[INFO] [stderr]   cargo:rerun-if-env-changed=CXXFLAGS_x86_64-unknown-linux-gnu
[INFO] [stderr]   CXXFLAGS_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CXXFLAGS_x86_64_unknown_linux_gnu
[INFO] [stderr]   CXXFLAGS_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_CXXFLAGS
[INFO] [stderr]   HOST_CXXFLAGS = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CXXFLAGS
[INFO] [stderr]   CXXFLAGS = None
[INFO] [stderr]   cargo:warning=./binding.cpp: In function 'int get_embeddings(void*, void*, float*)':
[INFO] [stderr]   cargo:warning=./binding.cpp:80:23: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=   80 |         if (llama_eval(ctx, embd_inp.data(), embd_inp.size(), n_past))
[INFO] [stderr]   cargo:warning=      |             ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr]   cargo:warning=  423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr]   cargo:warning=      |                              ^~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp: In function 'int eval(void*, void*, char*)':
[INFO] [stderr]   cargo:warning=./binding.cpp:138:22: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  138 |     return llama_eval(ctx, tokens.data(), n_prompt_tokens, n_past);
[INFO] [stderr]   cargo:warning=      |            ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr]   cargo:warning=  423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr]   cargo:warning=      |                              ^~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp: In function 'int llama_predict(void*, void*, char*, bool)':
[INFO] [stderr]   cargo:warning=./binding.cpp:282:19: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  282 |         llama_eval(ctx, tmp, 1, 0);
[INFO] [stderr]   cargo:warning=      |         ~~~~~~~~~~^~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr]   cargo:warning=  423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr]   cargo:warning=      |                              ^~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp:353:31: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  353 |                 if (llama_eval(ctx, &embd[i], n_eval, n_past))
[INFO] [stderr]   cargo:warning=      |                     ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr]   cargo:warning=  423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr]   cargo:warning=      |                              ^~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp:440:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  440 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr]   cargo:warning=      |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr]   cargo:warning=  621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr]   cargo:warning=      |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp:446:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  446 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr]   cargo:warning=      |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr]   cargo:warning=  621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr]   cargo:warning=      |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp:456:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  456 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr]   cargo:warning=      |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr]   cargo:warning=  621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr]   cargo:warning=      |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp:475:42: warning: cast from type 'const char*' to type 'char*' casts away qualifiers [-Wcast-qual]
[INFO] [stderr]   cargo:warning=  475 |             if (!tokenCallback(state_pr, (char*)token_str.c_str()))
[INFO] [stderr]   cargo:warning=      |                                          ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./binding.cpp: In function 'void* llama_allocate_params(const char*, int, int, int, int, float, float, float, int, bool, bool, int, int, const char**, int, float, float, float, float, int, float, float, bool, const char*, const char*, bool, bool, bool, const char*, const char*, bool)':
[INFO] [stderr]   cargo:warning=./binding.cpp:629:100: warning: unused parameter 'ignore_eos' [-Wunused-parameter]
[INFO] [stderr]   cargo:warning=  629 |                             float top_p, float temp, float repeat_penalty, int repeat_last_n, bool ignore_eos, bool memory_f16, int n_batch, int n_keep, const char **antiprompt, int antiprompt_count,
[INFO] [stderr]   cargo:warning=      |                                                                                               ~~~~~^~~~~~~~~~
[INFO] [stderr]   cargo:warning=./binding.cpp: In function 'void* load_model(const char*, int, int, bool, bool, bool, bool, bool, bool, int, int, const char*, const char*, bool)':
[INFO] [stderr]   cargo:warning=./binding.cpp:708:122: warning: unused parameter 'low_vram' [-Wunused-parameter]
[INFO] [stderr]   cargo:warning=  708 | void *load_model(const char *fname, int n_ctx, int n_seed, bool memory_f16, bool mlock, bool embeddings, bool mmap, bool low_vram, bool vocab_only, int n_gpu_layers, int n_batch, const char *maingpu, const char *tensorsplit, bool numa)
[INFO] [stderr]   cargo:warning=      |                                                                                                                     ~~~~~^~~~~~~~
[INFO] [stderr]   cargo:rerun-if-env-changed=AR_x86_64-unknown-linux-gnu
[INFO] [stderr]   AR_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=AR_x86_64_unknown_linux_gnu
[INFO] [stderr]   AR_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_AR
[INFO] [stderr]   HOST_AR = None
[INFO] [stderr]   cargo:rerun-if-env-changed=AR
[INFO] [stderr]   AR = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS_x86_64-unknown-linux-gnu
[INFO] [stderr]   ARFLAGS_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS_x86_64_unknown_linux_gnu
[INFO] [stderr]   ARFLAGS_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_ARFLAGS
[INFO] [stderr]   HOST_ARFLAGS = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS
[INFO] [stderr]   ARFLAGS = None
[INFO] [stderr]   cargo:warning=ar: /opt/rustwide/target/debug/build/llama_cpp_rs-a791689fede158f0/out/llama.cpp/ggml.o: No such file or directory
[INFO] [stderr] 
[INFO] [stderr]   --- stderr
[INFO] [stderr]   error: 'rustfmt' is not installed for the custom toolchain '30f74ff0dc4d66debc8b50724c446f817e5f75f4'.
[INFO] [stderr]   note: this is a custom toolchain, which cannot use `rustup component add`
[INFO] [stderr]   help: if you built this toolchain from source, and used `rustup toolchain link`, then you may be able to build the component with `x.py`
[INFO] [stderr]   Failed to run rustfmt: Internal rustfmt error (non-fatal, continuing)
[INFO] [stderr] 
[INFO] [stderr] 
[INFO] [stderr]   error occurred: Command ZERO_AR_DATE="1" "ar" "cq" "/opt/rustwide/target/debug/build/llama_cpp_rs-a791689fede158f0/out/libbinding.a" "/opt/rustwide/target/debug/build/llama_cpp_rs-a791689fede158f0/out/073db387043af495-common.o" "/opt/rustwide/target/debug/build/llama_cpp_rs-a791689fede158f0/out/30b5508d68fcb5a8-llama.o" "/opt/rustwide/target/debug/build/llama_cpp_rs-a791689fede158f0/out/8f1a5a601f45df90-binding.o" "/opt/rustwide/target/debug/build/llama_cpp_rs-a791689fede158f0/out/llama.cpp/ggml.o" with args "ar" did not execute successfully (status code exit status: 1).
[INFO] [stderr] 
[INFO] [stderr] 
[INFO] running `Command { std: "docker" "inspect" "668684f68d7d56c1759a49a4780e1e55279a2bafa338c71e0452202f47e0c601", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "668684f68d7d56c1759a49a4780e1e55279a2bafa338c71e0452202f47e0c601", kill_on_drop: false }`
[INFO] [stdout] 668684f68d7d56c1759a49a4780e1e55279a2bafa338c71e0452202f47e0c601
[INFO] checking llama_cpp_rs-0.3.0 against try#72eda894eb3548c7ba774079ce0afab42958d4ee for pr-121848-1
[INFO] extracting crate llama_cpp_rs 0.3.0 into /workspace/builds/worker-4-tc2/source
[INFO] validating manifest of crates.io crate llama_cpp_rs 0.3.0 on toolchain 72eda894eb3548c7ba774079ce0afab42958d4ee
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+72eda894eb3548c7ba774079ce0afab42958d4ee" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }`
[INFO] started tweaking crates.io crate llama_cpp_rs 0.3.0
[INFO] finished tweaking crates.io crate llama_cpp_rs 0.3.0
[INFO] tweaked toml for crates.io crate llama_cpp_rs 0.3.0 written to /workspace/builds/worker-4-tc2/source/Cargo.toml
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+72eda894eb3548c7ba774079ce0afab42958d4ee" "generate-lockfile" "--manifest-path" "Cargo.toml", kill_on_drop: false }`
[INFO] [stderr]     Blocking waiting for file lock on package cache
[INFO] [stderr]     Updating crates.io index
[INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+72eda894eb3548c7ba774079ce0afab42958d4ee" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-4-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-4-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:2788e3201cd34a07e3172128adcd8b3090168a8e3bcc40d7c032b9dda1df7d1c" "/opt/rustwide/cargo-home/bin/cargo" "+72eda894eb3548c7ba774079ce0afab42958d4ee" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }`
[INFO] [stdout] 3ba3c1eb6f2d5ff67bb7fea71ad61ef00d629bf888d92a7182b39fb1a42e7923
[INFO] running `Command { std: "docker" "start" "-a" "3ba3c1eb6f2d5ff67bb7fea71ad61ef00d629bf888d92a7182b39fb1a42e7923", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "inspect" "3ba3c1eb6f2d5ff67bb7fea71ad61ef00d629bf888d92a7182b39fb1a42e7923", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "3ba3c1eb6f2d5ff67bb7fea71ad61ef00d629bf888d92a7182b39fb1a42e7923", kill_on_drop: false }`
[INFO] [stdout] 3ba3c1eb6f2d5ff67bb7fea71ad61ef00d629bf888d92a7182b39fb1a42e7923
[INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-4-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-4-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:2788e3201cd34a07e3172128adcd8b3090168a8e3bcc40d7c032b9dda1df7d1c" "/opt/rustwide/cargo-home/bin/cargo" "+72eda894eb3548c7ba774079ce0afab42958d4ee" "check" "--frozen" "--all" "--all-targets" "--message-format=json", kill_on_drop: false }`
[INFO] [stdout] 888c06c3575b8aee6955b3bebf7ad00cf3d866d4d014b69bb2ee696808b9ff20
[INFO] running `Command { std: "docker" "start" "-a" "888c06c3575b8aee6955b3bebf7ad00cf3d866d4d014b69bb2ee696808b9ff20", kill_on_drop: false }`
[INFO] [stderr]    Compiling glob v0.3.1
[INFO] [stderr]    Compiling prettyplease v0.2.17
[INFO] [stderr]    Compiling proc-macro2 v1.0.79
[INFO] [stderr]    Compiling rustix v0.38.32
[INFO] [stderr]    Compiling minimal-lexical v0.2.1
[INFO] [stderr]    Compiling bitflags v2.5.0
[INFO] [stderr]    Compiling regex-syntax v0.8.3
[INFO] [stderr]    Compiling libc v0.2.153
[INFO] [stderr]    Compiling linux-raw-sys v0.4.13
[INFO] [stderr]    Compiling libloading v0.8.3
[INFO] [stderr]    Compiling nom v7.1.3
[INFO] [stderr]    Compiling quote v1.0.35
[INFO] [stderr]    Compiling clang-sys v1.7.0
[INFO] [stderr]    Compiling bindgen v0.66.1
[INFO] [stderr]    Compiling syn v2.0.57
[INFO] [stderr]    Compiling home v0.5.9
[INFO] [stderr]    Compiling either v1.10.0
[INFO] [stderr]    Compiling lazy_static v1.4.0
[INFO] [stderr]    Compiling peeking_take_while v0.1.2
[INFO] [stderr]    Compiling shlex v1.3.0
[INFO] [stderr]    Compiling lazycell v1.3.0
[INFO] [stderr]    Compiling rustc-hash v1.1.0
[INFO] [stderr]    Compiling regex-automata v0.4.6
[INFO] [stderr]    Compiling cexpr v0.6.0
[INFO] [stderr]    Compiling which v4.4.2
[INFO] [stderr]    Compiling regex v1.10.4
[INFO] [stderr]    Compiling llama_cpp_rs v0.3.0 (/opt/rustwide/workdir)
[INFO] [stderr] The following warnings were emitted during compilation:
[INFO] [stderr] 
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: Compiler version doesn't include clang or GCC: "cc" "--version"
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/ggml.c:19672:13: warning: 'ggml_opt_get_grad' defined but not used [-Wunused-function]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: 19672 | static void ggml_opt_get_grad(int np, struct ggml_tensor * const ps[], float * g) {
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |             ^~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/k_quants.c:205:14: warning: 'make_qkx1_quants' defined but not used [-Wunused-function]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   205 | static float make_qkx1_quants(int n, int nmax, const float * restrict x, uint8_t * restrict L, float * restrict the_min,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |              ^~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: Compiler version doesn't include clang or GCC: "c++" "--version"
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'int get_embeddings(void*, void*, float*)':
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:80:23: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    80 |         if (llama_eval(ctx, embd_inp.data(), embd_inp.size(), n_past))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |             ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                              ^~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'int eval(void*, void*, char*)':
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:138:22: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   138 |     return llama_eval(ctx, tokens.data(), n_prompt_tokens, n_past);
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |            ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                              ^~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'int llama_predict(void*, void*, char*, bool)':
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:282:19: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   282 |         llama_eval(ctx, tmp, 1, 0);
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |         ~~~~~~~~~~^~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                              ^~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:353:31: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   353 |                 if (llama_eval(ctx, &embd[i], n_eval, n_past))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                     ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                              ^~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:440:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   440 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:446:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   446 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:456:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   456 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:                  from ./binding.cpp:1:
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:    31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                    ^~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:475:42: warning: cast from type 'const char*' to type 'char*' casts away qualifiers [-Wcast-qual]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   475 |             if (!tokenCallback(state_pr, (char*)token_str.c_str()))
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                          ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'void* llama_allocate_params(const char*, int, int, int, int, float, float, float, int, bool, bool, int, int, const char**, int, float, float, float, float, int, float, float, bool, const char*, const char*, bool, bool, bool, const char*, const char*, bool)':
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:629:100: warning: unused parameter 'ignore_eos' [-Wunused-parameter]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   629 |                             float top_p, float temp, float repeat_penalty, int repeat_last_n, bool ignore_eos, bool memory_f16, int n_batch, int n_keep, const char **antiprompt, int antiprompt_count,
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                                                                               ~~~~~^~~~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp: In function 'void* load_model(const char*, int, int, bool, bool, bool, bool, bool, bool, int, int, const char*, const char*, bool)':
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ./binding.cpp:708:122: warning: unused parameter 'low_vram' [-Wunused-parameter]
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:   708 | void *load_model(const char *fname, int n_ctx, int n_seed, bool memory_f16, bool mlock, bool embeddings, bool mmap, bool low_vram, bool vocab_only, int n_gpu_layers, int n_batch, const char *maingpu, const char *tensorsplit, bool numa)
[INFO] [stderr] warning: llama_cpp_rs@0.3.0:       |                                                                                                                     ~~~~~^~~~~~~~
[INFO] [stderr] warning: llama_cpp_rs@0.3.0: ar: /opt/rustwide/target/debug/build/llama_cpp_rs-a791689fede158f0/out/llama.cpp/ggml.o: No such file or directory
[INFO] [stderr] 
[INFO] [stderr] error: failed to run custom build command for `llama_cpp_rs v0.3.0 (/opt/rustwide/workdir)`
[INFO] [stderr] note: To improve backtraces for build dependencies, set the CARGO_PROFILE_DEV_BUILD_OVERRIDE_DEBUG=true environment variable to enable debug information generation.
[INFO] [stderr] 
[INFO] [stderr] Caused by:
[INFO] [stderr]   process didn't exit successfully: `/opt/rustwide/target/debug/build/llama_cpp_rs-3d7548e85adc98be/build-script-build` (exit status: 1)
[INFO] [stderr]   --- stdout
[INFO] [stderr]   cargo:rerun-if-env-changed=TARGET
[INFO] [stderr]   cargo:rerun-if-env-changed=BINDGEN_EXTRA_CLANG_ARGS_x86_64-unknown-linux-gnu
[INFO] [stderr]   cargo:rerun-if-env-changed=BINDGEN_EXTRA_CLANG_ARGS_x86_64_unknown_linux_gnu
[INFO] [stderr]   cargo:rerun-if-env-changed=BINDGEN_EXTRA_CLANG_ARGS
[INFO] [stderr]   cargo:rerun-if-changed=/usr/lib/llvm-14/lib/clang/14.0.0/include/stdbool.h
[INFO] [stderr]   TARGET = Some("x86_64-unknown-linux-gnu")
[INFO] [stderr]   OPT_LEVEL = Some("0")
[INFO] [stderr]   HOST = Some("x86_64-unknown-linux-gnu")
[INFO] [stderr]   cargo:rerun-if-env-changed=CC_x86_64-unknown-linux-gnu
[INFO] [stderr]   CC_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CC_x86_64_unknown_linux_gnu
[INFO] [stderr]   CC_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_CC
[INFO] [stderr]   HOST_CC = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CC
[INFO] [stderr]   CC = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CC_ENABLE_DEBUG_OUTPUT
[INFO] [stderr]   cargo:warning=Compiler version doesn't include clang or GCC: "cc" "--version"
[INFO] [stderr]   cargo:rerun-if-env-changed=CRATE_CC_NO_DEFAULTS
[INFO] [stderr]   CRATE_CC_NO_DEFAULTS = None
[INFO] [stderr]   DEBUG = Some("true")
[INFO] [stderr]   CARGO_CFG_TARGET_FEATURE = Some("fxsr,sse,sse2")
[INFO] [stderr]   cargo:rerun-if-env-changed=CFLAGS_x86_64-unknown-linux-gnu
[INFO] [stderr]   CFLAGS_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CFLAGS_x86_64_unknown_linux_gnu
[INFO] [stderr]   CFLAGS_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_CFLAGS
[INFO] [stderr]   HOST_CFLAGS = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CFLAGS
[INFO] [stderr]   CFLAGS = None
[INFO] [stderr]   cargo:warning=./llama.cpp/ggml.c:19672:13: warning: 'ggml_opt_get_grad' defined but not used [-Wunused-function]
[INFO] [stderr]   cargo:warning=19672 | static void ggml_opt_get_grad(int np, struct ggml_tensor * const ps[], float * g) {
[INFO] [stderr]   cargo:warning=      |             ^~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/k_quants.c:205:14: warning: 'make_qkx1_quants' defined but not used [-Wunused-function]
[INFO] [stderr]   cargo:warning=  205 | static float make_qkx1_quants(int n, int nmax, const float * restrict x, uint8_t * restrict L, float * restrict the_min,
[INFO] [stderr]   cargo:warning=      |              ^~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:rerun-if-env-changed=AR_x86_64-unknown-linux-gnu
[INFO] [stderr]   AR_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=AR_x86_64_unknown_linux_gnu
[INFO] [stderr]   AR_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_AR
[INFO] [stderr]   HOST_AR = None
[INFO] [stderr]   cargo:rerun-if-env-changed=AR
[INFO] [stderr]   AR = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS_x86_64-unknown-linux-gnu
[INFO] [stderr]   ARFLAGS_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS_x86_64_unknown_linux_gnu
[INFO] [stderr]   ARFLAGS_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_ARFLAGS
[INFO] [stderr]   HOST_ARFLAGS = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS
[INFO] [stderr]   ARFLAGS = None
[INFO] [stderr]   cargo:rustc-link-lib=static=ggml
[INFO] [stderr]   cargo:rustc-link-search=native=/opt/rustwide/target/debug/build/llama_cpp_rs-a791689fede158f0/out
[INFO] [stderr]   TARGET = Some("x86_64-unknown-linux-gnu")
[INFO] [stderr]   OPT_LEVEL = Some("0")
[INFO] [stderr]   HOST = Some("x86_64-unknown-linux-gnu")
[INFO] [stderr]   cargo:rerun-if-env-changed=CXX_x86_64-unknown-linux-gnu
[INFO] [stderr]   CXX_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CXX_x86_64_unknown_linux_gnu
[INFO] [stderr]   CXX_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_CXX
[INFO] [stderr]   HOST_CXX = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CXX
[INFO] [stderr]   CXX = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CC_ENABLE_DEBUG_OUTPUT
[INFO] [stderr]   cargo:warning=Compiler version doesn't include clang or GCC: "c++" "--version"
[INFO] [stderr]   cargo:rerun-if-env-changed=CRATE_CC_NO_DEFAULTS
[INFO] [stderr]   CRATE_CC_NO_DEFAULTS = None
[INFO] [stderr]   DEBUG = Some("true")
[INFO] [stderr]   CARGO_CFG_TARGET_FEATURE = Some("fxsr,sse,sse2")
[INFO] [stderr]   cargo:rerun-if-env-changed=CXXFLAGS_x86_64-unknown-linux-gnu
[INFO] [stderr]   CXXFLAGS_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CXXFLAGS_x86_64_unknown_linux_gnu
[INFO] [stderr]   CXXFLAGS_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_CXXFLAGS
[INFO] [stderr]   HOST_CXXFLAGS = None
[INFO] [stderr]   cargo:rerun-if-env-changed=CXXFLAGS
[INFO] [stderr]   CXXFLAGS = None
[INFO] [stderr]   cargo:warning=./binding.cpp: In function 'int get_embeddings(void*, void*, float*)':
[INFO] [stderr]   cargo:warning=./binding.cpp:80:23: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=   80 |         if (llama_eval(ctx, embd_inp.data(), embd_inp.size(), n_past))
[INFO] [stderr]   cargo:warning=      |             ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr]   cargo:warning=  423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr]   cargo:warning=      |                              ^~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp: In function 'int eval(void*, void*, char*)':
[INFO] [stderr]   cargo:warning=./binding.cpp:138:22: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  138 |     return llama_eval(ctx, tokens.data(), n_prompt_tokens, n_past);
[INFO] [stderr]   cargo:warning=      |            ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr]   cargo:warning=  423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr]   cargo:warning=      |                              ^~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp: In function 'int llama_predict(void*, void*, char*, bool)':
[INFO] [stderr]   cargo:warning=./binding.cpp:282:19: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  282 |         llama_eval(ctx, tmp, 1, 0);
[INFO] [stderr]   cargo:warning=      |         ~~~~~~~~~~^~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr]   cargo:warning=  423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr]   cargo:warning=      |                              ^~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp:353:31: warning: 'int llama_eval(llama_context*, llama_token*, int32_t, int)' is deprecated: use llama_decode() instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  353 |                 if (llama_eval(ctx, &embd[i], n_eval, n_past))
[INFO] [stderr]   cargo:warning=      |                     ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:423:30: note: declared here
[INFO] [stderr]   cargo:warning=  423 |     LLAMA_API DEPRECATED(int llama_eval(
[INFO] [stderr]   cargo:warning=      |                              ^~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp:440:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  440 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr]   cargo:warning=      |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr]   cargo:warning=  621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr]   cargo:warning=      |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp:446:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  446 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr]   cargo:warning=      |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr]   cargo:warning=  621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr]   cargo:warning=      |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp:456:49: warning: 'void llama_sample_temperature(llama_context*, llama_token_data_array*, float)' is deprecated: use llama_sample_temp instead [-Wdeprecated-declarations]
[INFO] [stderr]   cargo:warning=  456 |                         llama_sample_temperature(ctx, &candidates_p, temp);
[INFO] [stderr]   cargo:warning=      |                         ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=In file included from ./llama.cpp/common/common.h:5,
[INFO] [stderr]   cargo:warning=                 from ./binding.cpp:1:
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:621:31: note: declared here
[INFO] [stderr]   cargo:warning=  621 |     LLAMA_API DEPRECATED(void llama_sample_temperature(
[INFO] [stderr]   cargo:warning=      |                               ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./llama.cpp/llama.h:31:36: note: in definition of macro 'DEPRECATED'
[INFO] [stderr]   cargo:warning=   31 | #    define DEPRECATED(func, hint) func __attribute__((deprecated(hint)))
[INFO] [stderr]   cargo:warning=      |                                    ^~~~
[INFO] [stderr]   cargo:warning=./binding.cpp:475:42: warning: cast from type 'const char*' to type 'char*' casts away qualifiers [-Wcast-qual]
[INFO] [stderr]   cargo:warning=  475 |             if (!tokenCallback(state_pr, (char*)token_str.c_str()))
[INFO] [stderr]   cargo:warning=      |                                          ^~~~~~~~~~~~~~~~~~~~~~~~
[INFO] [stderr]   cargo:warning=./binding.cpp: In function 'void* llama_allocate_params(const char*, int, int, int, int, float, float, float, int, bool, bool, int, int, const char**, int, float, float, float, float, int, float, float, bool, const char*, const char*, bool, bool, bool, const char*, const char*, bool)':
[INFO] [stderr]   cargo:warning=./binding.cpp:629:100: warning: unused parameter 'ignore_eos' [-Wunused-parameter]
[INFO] [stderr]   cargo:warning=  629 |                             float top_p, float temp, float repeat_penalty, int repeat_last_n, bool ignore_eos, bool memory_f16, int n_batch, int n_keep, const char **antiprompt, int antiprompt_count,
[INFO] [stderr]   cargo:warning=      |                                                                                               ~~~~~^~~~~~~~~~
[INFO] [stderr]   cargo:warning=./binding.cpp: In function 'void* load_model(const char*, int, int, bool, bool, bool, bool, bool, bool, int, int, const char*, const char*, bool)':
[INFO] [stderr]   cargo:warning=./binding.cpp:708:122: warning: unused parameter 'low_vram' [-Wunused-parameter]
[INFO] [stderr]   cargo:warning=  708 | void *load_model(const char *fname, int n_ctx, int n_seed, bool memory_f16, bool mlock, bool embeddings, bool mmap, bool low_vram, bool vocab_only, int n_gpu_layers, int n_batch, const char *maingpu, const char *tensorsplit, bool numa)
[INFO] [stderr]   cargo:warning=      |                                                                                                                     ~~~~~^~~~~~~~
[INFO] [stderr]   cargo:rerun-if-env-changed=AR_x86_64-unknown-linux-gnu
[INFO] [stderr]   AR_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=AR_x86_64_unknown_linux_gnu
[INFO] [stderr]   AR_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_AR
[INFO] [stderr]   HOST_AR = None
[INFO] [stderr]   cargo:rerun-if-env-changed=AR
[INFO] [stderr]   AR = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS_x86_64-unknown-linux-gnu
[INFO] [stderr]   ARFLAGS_x86_64-unknown-linux-gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS_x86_64_unknown_linux_gnu
[INFO] [stderr]   ARFLAGS_x86_64_unknown_linux_gnu = None
[INFO] [stderr]   cargo:rerun-if-env-changed=HOST_ARFLAGS
[INFO] [stderr]   HOST_ARFLAGS = None
[INFO] [stderr]   cargo:rerun-if-env-changed=ARFLAGS
[INFO] [stderr]   ARFLAGS = None
[INFO] [stderr]   cargo:warning=ar: /opt/rustwide/target/debug/build/llama_cpp_rs-a791689fede158f0/out/llama.cpp/ggml.o: No such file or directory
[INFO] [stderr] 
[INFO] [stderr]   --- stderr
[INFO] [stderr]   error: 'rustfmt' is not installed for the custom toolchain '72eda894eb3548c7ba774079ce0afab42958d4ee'.
[INFO] [stderr]   note: this is a custom toolchain, which cannot use `rustup component add`
[INFO] [stderr]   help: if you built this toolchain from source, and used `rustup toolchain link`, then you may be able to build the component with `x.py`
[INFO] [stderr]   Failed to run rustfmt: Internal rustfmt error (non-fatal, continuing)
[INFO] [stderr] 
[INFO] [stderr] 
[INFO] [stderr]   error occurred: Command ZERO_AR_DATE="1" "ar" "cq" "/opt/rustwide/target/debug/build/llama_cpp_rs-a791689fede158f0/out/libbinding.a" "/opt/rustwide/target/debug/build/llama_cpp_rs-a791689fede158f0/out/073db387043af495-common.o" "/opt/rustwide/target/debug/build/llama_cpp_rs-a791689fede158f0/out/30b5508d68fcb5a8-llama.o" "/opt/rustwide/target/debug/build/llama_cpp_rs-a791689fede158f0/out/8f1a5a601f45df90-binding.o" "/opt/rustwide/target/debug/build/llama_cpp_rs-a791689fede158f0/out/llama.cpp/ggml.o" with args "ar" did not execute successfully (status code exit status: 1).
[INFO] [stderr] 
[INFO] [stderr] 
[INFO] running `Command { std: "docker" "inspect" "888c06c3575b8aee6955b3bebf7ad00cf3d866d4d014b69bb2ee696808b9ff20", kill_on_drop: false }`
[INFO] running `Command { std: "docker" "rm" "-f" "888c06c3575b8aee6955b3bebf7ad00cf3d866d4d014b69bb2ee696808b9ff20", kill_on_drop: false }`
[INFO] [stdout] 888c06c3575b8aee6955b3bebf7ad00cf3d866d4d014b69bb2ee696808b9ff20
