[INFO] cloning repository https://github.com/abughalib/llama_rust [INFO] running `Command { std: "git" "-c" "credential.helper=" "-c" "credential.helper=/workspace/cargo-home/bin/git-credential-null" "clone" "--bare" "https://github.com/abughalib/llama_rust" "/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2Fabughalib%2Fllama_rust", kill_on_drop: false }` [INFO] [stderr] Cloning into bare repository '/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2Fabughalib%2Fllama_rust'... [INFO] running `Command { std: "git" "rev-parse" "HEAD", kill_on_drop: false }` [INFO] [stdout] 9b0a11c0809bace1355f8c899429adf4bd5a71d8 [INFO] checking abughalib/llama_rust against master#cccf075eba88363269e8589ebb8d40874cc542d8 for pr-142681 [INFO] running `Command { std: "git" "clone" "/workspace/cache/git-repos/https%3A%2F%2Fgithub.com%2Fabughalib%2Fllama_rust" "/workspace/builds/worker-2-tc1/source", kill_on_drop: false }` [INFO] [stderr] Cloning into '/workspace/builds/worker-2-tc1/source'... [INFO] [stderr] done. [INFO] started tweaking git repo https://github.com/abughalib/llama_rust [INFO] finished tweaking git repo https://github.com/abughalib/llama_rust [INFO] tweaked toml for git repo https://github.com/abughalib/llama_rust written to /workspace/builds/worker-2-tc1/source/Cargo.toml [INFO] validating manifest of git repo https://github.com/abughalib/llama_rust on toolchain cccf075eba88363269e8589ebb8d40874cc542d8 [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+cccf075eba88363269e8589ebb8d40874cc542d8" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }` [INFO] crate git repo https://github.com/abughalib/llama_rust already has a lockfile, it will not be regenerated [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+cccf075eba88363269e8589ebb8d40874cc542d8" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }` [INFO] [stderr] Updating crates.io index [INFO] [stderr] Blocking waiting for file lock on package cache [INFO] [stderr] Downloading crates ... [INFO] [stderr] Downloaded llama_cpp v0.3.1 [INFO] [stderr] Downloaded llama_cpp_sys v0.3.1 [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-2-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-2-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:cf8efcab8866f2cf4285301c7418e2f4f2a9b088c91ba69c40d5b659f81557f7" "/opt/rustwide/cargo-home/bin/cargo" "+cccf075eba88363269e8589ebb8d40874cc542d8" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }` [INFO] [stdout] 708afbbc3f0660a6fbc709d77c41df7d4da78500a2aeaaf9984af257157c6716 [INFO] running `Command { std: "docker" "start" "-a" "708afbbc3f0660a6fbc709d77c41df7d4da78500a2aeaaf9984af257157c6716", kill_on_drop: false }` [INFO] running `Command { std: "docker" "inspect" "708afbbc3f0660a6fbc709d77c41df7d4da78500a2aeaaf9984af257157c6716", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "708afbbc3f0660a6fbc709d77c41df7d4da78500a2aeaaf9984af257157c6716", kill_on_drop: false }` [INFO] [stdout] 708afbbc3f0660a6fbc709d77c41df7d4da78500a2aeaaf9984af257157c6716 [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-2-tc1/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-2-tc1/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:cf8efcab8866f2cf4285301c7418e2f4f2a9b088c91ba69c40d5b659f81557f7" "/opt/rustwide/cargo-home/bin/cargo" "+cccf075eba88363269e8589ebb8d40874cc542d8" "check" "--frozen" "--all" "--all-targets" "--message-format=json", kill_on_drop: false }` [INFO] [stdout] fe0ce4980f06865ca7ca39d8f48d47b78b507e319e6ee6007a4991f0440c67cc [INFO] running `Command { std: "docker" "start" "-a" "fe0ce4980f06865ca7ca39d8f48d47b78b507e319e6ee6007a4991f0440c67cc", kill_on_drop: false }` [INFO] [stderr] Compiling proc-macro2 v1.0.79 [INFO] [stderr] Compiling unicode-ident v1.0.12 [INFO] [stderr] Compiling libc v0.2.153 [INFO] [stderr] Compiling glob v0.3.1 [INFO] [stderr] Compiling rustix v0.38.32 [INFO] [stderr] Compiling prettyplease v0.2.17 [INFO] [stderr] Compiling bitflags v2.5.0 [INFO] [stderr] Compiling minimal-lexical v0.2.1 [INFO] [stderr] Compiling either v1.10.0 [INFO] [stderr] Compiling memchr v2.7.2 [INFO] [stderr] Compiling cfg-if v1.0.0 [INFO] [stderr] Compiling autocfg v1.2.0 [INFO] [stderr] Compiling regex-syntax v0.8.3 [INFO] [stderr] Compiling linux-raw-sys v0.4.13 [INFO] [stderr] Compiling bindgen v0.69.4 [INFO] [stderr] Checking pin-project-lite v0.2.14 [INFO] [stderr] Compiling libloading v0.8.3 [INFO] [stderr] Compiling home v0.5.9 [INFO] [stderr] Compiling log v0.4.21 [INFO] [stderr] Compiling lazycell v1.3.0 [INFO] [stderr] Compiling shlex v1.3.0 [INFO] [stderr] Checking futures-sink v0.3.30 [INFO] [stderr] Compiling itertools v0.12.1 [INFO] [stderr] Checking futures-core v0.3.30 [INFO] [stderr] Compiling lazy_static v1.4.0 [INFO] [stderr] Compiling rustc-hash v1.1.0 [INFO] [stderr] Checking pin-utils v0.1.0 [INFO] [stderr] Checking futures-task v0.3.30 [INFO] [stderr] Checking futures-io v0.3.30 [INFO] [stderr] Compiling once_cell v1.19.0 [INFO] [stderr] Compiling ash v0.37.3+1.3.251 [INFO] [stderr] Compiling syn v1.0.109 [INFO] [stderr] Compiling thiserror v1.0.58 [INFO] [stderr] Checking futures-channel v0.3.30 [INFO] [stderr] Compiling convert_case v0.4.0 [INFO] [stderr] Compiling nom v7.1.3 [INFO] [stderr] Compiling clang-sys v1.7.0 [INFO] [stderr] Compiling slab v0.4.9 [INFO] [stderr] Compiling quote v1.0.35 [INFO] [stderr] Checking tracing-core v0.1.32 [INFO] [stderr] Checking tokio v1.37.0 [INFO] [stderr] Compiling syn v2.0.58 [INFO] [stderr] Compiling regex-automata v0.4.6 [INFO] [stderr] Compiling jobserver v0.1.28 [INFO] [stderr] Checking num_cpus v1.16.0 [INFO] [stderr] Compiling cc v1.0.90 [INFO] [stderr] Compiling which v4.4.2 [INFO] [stderr] Compiling regex v1.10.4 [INFO] [stderr] Compiling cexpr v0.6.0 [INFO] [stderr] Compiling link-cplusplus v1.0.9 [INFO] [stderr] Compiling derive_more v0.99.17 [INFO] [stderr] Compiling futures-macro v0.3.30 [INFO] [stderr] Compiling thiserror-impl v1.0.58 [INFO] [stderr] Compiling tracing-attributes v0.1.27 [INFO] [stderr] Checking futures-util v0.3.30 [INFO] [stderr] Checking tracing v0.1.40 [INFO] [stderr] Checking futures-executor v0.3.30 [INFO] [stderr] Checking futures v0.3.30 [INFO] [stderr] Compiling llama_cpp_sys v0.3.1 [INFO] [stderr] warning: llama_cpp_sys@0.3.1: Compiler version doesn't include clang or GCC: "c++" "--version" [INFO] [stderr] warning: llama_cpp_sys@0.3.1: ./thirdparty/llama.cpp/ggml-vulkan.cpp: In function 'void ggml_vk_buffer_copy(vk_buffer&, size_t, vk_buffer&, size_t, size_t)': [INFO] [stderr] warning: llama_cpp_sys@0.3.1: ./thirdparty/llama.cpp/ggml-vulkan.cpp:1938:22: warning: unused variable 'bc' [-Wunused-variable] [INFO] [stderr] warning: llama_cpp_sys@0.3.1: 1938 | VkBufferCopy bc{ src_offset, dst_offset, size }; [INFO] [stderr] warning: llama_cpp_sys@0.3.1: | ^~ [INFO] [stderr] warning: llama_cpp_sys@0.3.1: ./thirdparty/llama.cpp/ggml-vulkan.cpp: In function 'bool ggml_backend_vk_buffer_cpy_tensor(ggml_backend_buffer_t, const ggml_tensor*, ggml_tensor*)': [INFO] [stderr] warning: llama_cpp_sys@0.3.1: ./thirdparty/llama.cpp/ggml-vulkan.cpp:4784:42: warning: unused variable 'ctx' [-Wunused-variable] [INFO] [stderr] warning: llama_cpp_sys@0.3.1: 4784 | ggml_backend_vk_buffer_context * ctx = (ggml_backend_vk_buffer_context *)buffer->context; [INFO] [stderr] warning: llama_cpp_sys@0.3.1: | ^~~ [INFO] [stderr] warning: llama_cpp_sys@0.3.1: Compiler version doesn't include clang or GCC: "cc" "--version" [INFO] [stderr] warning: llama_cpp_sys@0.3.1: ./thirdparty/llama.cpp/ggml.c:150: warning: "GGML_DEBUG" redefined [INFO] [stderr] warning: llama_cpp_sys@0.3.1: 150 | #define GGML_DEBUG 0 [INFO] [stderr] warning: llama_cpp_sys@0.3.1: | [INFO] [stderr] warning: llama_cpp_sys@0.3.1: : note: this is the location of the previous definition [INFO] [stderr] warning: llama_cpp_sys@0.3.1: ./thirdparty/llama.cpp/ggml-alloc.c:448:13: warning: 'ggml_gallocr_set_node_offset' defined but not used [-Wunused-function] [INFO] [stderr] warning: llama_cpp_sys@0.3.1: 448 | static void ggml_gallocr_set_node_offset(ggml_gallocr_t galloc, struct ggml_tensor * node, int buffer_id, size_t offset) { [INFO] [stderr] warning: llama_cpp_sys@0.3.1: | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~ [INFO] [stderr] warning: llama_cpp_sys@0.3.1: ./thirdparty/llama.cpp/ggml-backend.c:1067:13: warning: 'ggml_backend_sched_print_assignments' defined but not used [-Wunused-function] [INFO] [stderr] warning: llama_cpp_sys@0.3.1: 1067 | static void ggml_backend_sched_print_assignments(ggml_backend_sched_t sched, struct ggml_cgraph * graph) { [INFO] [stderr] warning: llama_cpp_sys@0.3.1: | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ [INFO] [stderr] warning: llama_cpp_sys@0.3.1: ./thirdparty/llama.cpp/ggml-quants.c:1389:14: warning: 'make_qkx1_quants' defined but not used [-Wunused-function] [INFO] [stderr] warning: llama_cpp_sys@0.3.1: 1389 | static float make_qkx1_quants(int n, int nmax, const float * restrict x, uint8_t * restrict L, float * restrict the_min, [INFO] [stderr] warning: llama_cpp_sys@0.3.1: | ^~~~~~~~~~~~~~~~ [INFO] [stderr] warning: llama_cpp_sys@0.3.1: ./thirdparty/llama.cpp/ggml.c:18561:13: warning: 'ggml_opt_get_grad' defined but not used [-Wunused-function] [INFO] [stderr] warning: llama_cpp_sys@0.3.1: 18561 | static void ggml_opt_get_grad(int np, struct ggml_tensor * const ps[], float * g) { [INFO] [stderr] warning: llama_cpp_sys@0.3.1: | ^~~~~~~~~~~~~~~~~ [INFO] [stderr] warning: llama_cpp_sys@0.3.1: ./thirdparty/llama.cpp/llama.cpp: In function 'ggml_tensor* llm_build_kqv(ggml_context*, const llama_model&, const llama_hparams&, const llama_kv_cache&, ggml_cgraph*, ggml_tensor*, ggml_tensor*, ggml_tensor*, ggml_tensor*, ggml_tensor*, int64_t, int32_t, int32_t, float, const llm_build_cb&, int)': [INFO] [stderr] warning: llama_cpp_sys@0.3.1: ./thirdparty/llama.cpp/llama.cpp:4887:108: note: '#pragma message: TODO: ALiBi support in ggml_soft_max_ext is not implemented for Vulkan, Kompute, and SYCL' [INFO] [stderr] warning: llama_cpp_sys@0.3.1: 4887 | #pragma message("TODO: ALiBi support in ggml_soft_max_ext is not implemented for Vulkan, Kompute, and SYCL") [INFO] [stderr] warning: llama_cpp_sys@0.3.1: | ^ [INFO] [stderr] warning: llama_cpp_sys@0.3.1: ./thirdparty/llama.cpp/llama.cpp:4888:87: note: '#pragma message: Falling back to ggml_alibi(). Will become an error in Mar 2024' [INFO] [stderr] warning: llama_cpp_sys@0.3.1: 4888 | #pragma message(" Falling back to ggml_alibi(). Will become an error in Mar 2024") [INFO] [stderr] warning: llama_cpp_sys@0.3.1: | ^ [INFO] [stderr] warning: llama_cpp_sys@0.3.1: ./thirdparty/llama.cpp/llama.cpp:4889:73: note: '#pragma message: ref: https://github.com/ggerganov/llama.cpp/pull/5488' [INFO] [stderr] warning: llama_cpp_sys@0.3.1: 4889 | #pragma message("ref: https://github.com/ggerganov/llama.cpp/pull/5488") [INFO] [stderr] warning: llama_cpp_sys@0.3.1: | ^ [INFO] [stderr] warning: llama_cpp_sys@0.3.1: ./thirdparty/llama.cpp/llama.cpp:4894:24: warning: 'ggml_tensor* ggml_alibi(ggml_context*, ggml_tensor*, int, int, float)' is deprecated: use ggml_soft_max_ext instead (will be removed in Mar 2024) [-Wdeprecated-declarations] [INFO] [stderr] warning: llama_cpp_sys@0.3.1: 4894 | kq = ggml_alibi(ctx, kq, /*n_past*/ 0, n_head, hparams.f_max_alibi_bias); [INFO] [stderr] warning: llama_cpp_sys@0.3.1: | ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ [INFO] [stderr] warning: llama_cpp_sys@0.3.1: In file included from ./thirdparty/llama.cpp/llama.h:4, [INFO] [stderr] warning: llama_cpp_sys@0.3.1: from ./thirdparty/llama.cpp/llama.cpp:2: [INFO] [stderr] warning: llama_cpp_sys@0.3.1: ./thirdparty/llama.cpp/ggml.h:1502:51: note: declared here [INFO] [stderr] warning: llama_cpp_sys@0.3.1: 1502 | GGML_DEPRECATED(GGML_API struct ggml_tensor * ggml_alibi( [INFO] [stderr] warning: llama_cpp_sys@0.3.1: | ^~~~~~~~~~ [INFO] [stderr] warning: llama_cpp_sys@0.3.1: ./thirdparty/llama.cpp/ggml.h:202:41: note: in definition of macro 'GGML_DEPRECATED' [INFO] [stderr] warning: llama_cpp_sys@0.3.1: 202 | # define GGML_DEPRECATED(func, hint) func __attribute__((deprecated(hint))) [INFO] [stderr] warning: llama_cpp_sys@0.3.1: | ^~~~ [INFO] [stderr] warning: llama_cpp_sys@0.3.1: ./thirdparty/llama.cpp/llama.cpp: At global scope: [INFO] [stderr] warning: llama_cpp_sys@0.3.1: ./thirdparty/llama.cpp/llama.cpp:12940:1: fatal error: error writing to /tmp/cc3lMeL6.s: No space left on device [INFO] [stderr] warning: llama_cpp_sys@0.3.1: 12940 | } [INFO] [stderr] warning: llama_cpp_sys@0.3.1: | ^ [INFO] [stderr] warning: llama_cpp_sys@0.3.1: compilation terminated. [INFO] [stderr] error: failed to run custom build command for `llama_cpp_sys v0.3.1` [INFO] [stderr] note: To improve backtraces for build dependencies, set the CARGO_PROFILE_DEV_BUILD_OVERRIDE_DEBUG=true environment variable to enable debug information generation. [INFO] [stderr] [INFO] [stderr] Caused by: [INFO] [stderr] process didn't exit successfully: `/opt/rustwide/target/debug/build/llama_cpp_sys-36db343468810d4c/build-script-build` (exit status: 1) [INFO] [stderr] --- stdout [INFO] [stderr] cargo:rerun-if-changed=./thirdparty/llama.cpp [INFO] [stderr] Generating bindings.. [INFO] [stderr] cargo:rerun-if-env-changed=TARGET [INFO] [stderr] cargo:rerun-if-env-changed=BINDGEN_EXTRA_CLANG_ARGS_x86_64-unknown-linux-gnu [INFO] [stderr] cargo:rerun-if-env-changed=BINDGEN_EXTRA_CLANG_ARGS_x86_64_unknown_linux_gnu [INFO] [stderr] cargo:rerun-if-env-changed=BINDGEN_EXTRA_CLANG_ARGS [INFO] [stderr] cargo:rerun-if-changed=/usr/lib/llvm-18/lib/clang/18/include/stdint.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/stdint.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/libc-header-start.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/features.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/features-time64.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/wordsize.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/timesize.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/wordsize.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/stdc-predef.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/sys/cdefs.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/wordsize.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/long-double.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/gnu/stubs.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/gnu/stubs-64.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/types.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/features.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/wordsize.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/timesize.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/wordsize.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/typesizes.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/time64.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/wchar.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/wordsize.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/stdint-intn.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/types.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/stdint-uintn.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/types.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/stdint-least.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/types.h [INFO] [stderr] cargo:rerun-if-changed=/usr/lib/llvm-18/lib/clang/18/include/stddef.h [INFO] [stderr] cargo:rerun-if-changed=/usr/lib/llvm-18/lib/clang/18/include/__stddef_ptrdiff_t.h [INFO] [stderr] cargo:rerun-if-changed=/usr/lib/llvm-18/lib/clang/18/include/__stddef_size_t.h [INFO] [stderr] cargo:rerun-if-changed=/usr/lib/llvm-18/lib/clang/18/include/__stddef_wchar_t.h [INFO] [stderr] cargo:rerun-if-changed=/usr/lib/llvm-18/lib/clang/18/include/__stddef_null.h [INFO] [stderr] cargo:rerun-if-changed=/usr/lib/llvm-18/lib/clang/18/include/__stddef_max_align_t.h [INFO] [stderr] cargo:rerun-if-changed=/usr/lib/llvm-18/lib/clang/18/include/__stddef_offsetof.h [INFO] [stderr] cargo:rerun-if-changed=/usr/lib/llvm-18/lib/clang/18/include/stdbool.h [INFO] [stderr] cargo:rerun-if-changed=./thirdparty/llama.cpp/ggml.h [INFO] [stderr] cargo:rerun-if-changed=./thirdparty/llama.cpp/ggml-backend.h [INFO] [stderr] cargo:rerun-if-changed=./thirdparty/llama.cpp/ggml.h [INFO] [stderr] cargo:rerun-if-changed=./thirdparty/llama.cpp/ggml-alloc.h [INFO] [stderr] cargo:rerun-if-changed=./thirdparty/llama.cpp/ggml.h [INFO] [stderr] cargo:rerun-if-changed=/usr/lib/llvm-18/lib/clang/18/include/stddef.h [INFO] [stderr] cargo:rerun-if-changed=/usr/lib/llvm-18/lib/clang/18/include/stdint.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/stdio.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/libc-header-start.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/features.h [INFO] [stderr] cargo:rerun-if-changed=/usr/lib/llvm-18/lib/clang/18/include/stddef.h [INFO] [stderr] cargo:rerun-if-changed=/usr/lib/llvm-18/lib/clang/18/include/__stddef_size_t.h [INFO] [stderr] cargo:rerun-if-changed=/usr/lib/llvm-18/lib/clang/18/include/__stddef_null.h [INFO] [stderr] cargo:rerun-if-changed=/usr/lib/llvm-18/lib/clang/18/include/stdarg.h [INFO] [stderr] cargo:rerun-if-changed=/usr/lib/llvm-18/lib/clang/18/include/__stdarg___gnuc_va_list.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/types.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/types/__fpos_t.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/types.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/types/__mbstate_t.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/types/__fpos64_t.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/types.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/types/__mbstate_t.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/types/__FILE.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/types/FILE.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/types/struct_FILE.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/types.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/types/cookie_io_functions_t.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/types.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/stdio_lim.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/floatn.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/features.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/floatn-common.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/features.h [INFO] [stderr] cargo:rerun-if-changed=/usr/include/x86_64-linux-gnu/bits/long-double.h [INFO] [stderr] cargo:rerun-if-changed=/usr/lib/llvm-18/lib/clang/18/include/stdbool.h [INFO] [stderr] Compiling Vulkan GGML.. [INFO] [stderr] TARGET = Some("x86_64-unknown-linux-gnu") [INFO] [stderr] OPT_LEVEL = Some("0") [INFO] [stderr] HOST = Some("x86_64-unknown-linux-gnu") [INFO] [stderr] cargo:rerun-if-env-changed=CXX_x86_64-unknown-linux-gnu [INFO] [stderr] CXX_x86_64-unknown-linux-gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=CXX_x86_64_unknown_linux_gnu [INFO] [stderr] CXX_x86_64_unknown_linux_gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=HOST_CXX [INFO] [stderr] HOST_CXX = None [INFO] [stderr] cargo:rerun-if-env-changed=CXX [INFO] [stderr] CXX = None [INFO] [stderr] cargo:rerun-if-env-changed=CC_ENABLE_DEBUG_OUTPUT [INFO] [stderr] cargo:warning=Compiler version doesn't include clang or GCC: "c++" "--version" [INFO] [stderr] cargo:rerun-if-env-changed=CRATE_CC_NO_DEFAULTS [INFO] [stderr] CRATE_CC_NO_DEFAULTS = None [INFO] [stderr] DEBUG = Some("true") [INFO] [stderr] cargo:rerun-if-env-changed=CXXFLAGS_x86_64-unknown-linux-gnu [INFO] [stderr] CXXFLAGS_x86_64-unknown-linux-gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=CXXFLAGS_x86_64_unknown_linux_gnu [INFO] [stderr] CXXFLAGS_x86_64_unknown_linux_gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=HOST_CXXFLAGS [INFO] [stderr] HOST_CXXFLAGS = None [INFO] [stderr] cargo:rerun-if-env-changed=CXXFLAGS [INFO] [stderr] CXXFLAGS = None [INFO] [stderr] cargo:warning=./thirdparty/llama.cpp/ggml-vulkan.cpp: In function 'void ggml_vk_buffer_copy(vk_buffer&, size_t, vk_buffer&, size_t, size_t)': [INFO] [stderr] cargo:warning=./thirdparty/llama.cpp/ggml-vulkan.cpp:1938:22: warning: unused variable 'bc' [-Wunused-variable] [INFO] [stderr] cargo:warning= 1938 | VkBufferCopy bc{ src_offset, dst_offset, size }; [INFO] [stderr] cargo:warning= | ^~ [INFO] [stderr] cargo:warning=./thirdparty/llama.cpp/ggml-vulkan.cpp: In function 'bool ggml_backend_vk_buffer_cpy_tensor(ggml_backend_buffer_t, const ggml_tensor*, ggml_tensor*)': [INFO] [stderr] cargo:warning=./thirdparty/llama.cpp/ggml-vulkan.cpp:4784:42: warning: unused variable 'ctx' [-Wunused-variable] [INFO] [stderr] cargo:warning= 4784 | ggml_backend_vk_buffer_context * ctx = (ggml_backend_vk_buffer_context *)buffer->context; [INFO] [stderr] cargo:warning= | ^~~ [INFO] [stderr] cargo:rerun-if-env-changed=AR_x86_64-unknown-linux-gnu [INFO] [stderr] AR_x86_64-unknown-linux-gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=AR_x86_64_unknown_linux_gnu [INFO] [stderr] AR_x86_64_unknown_linux_gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=HOST_AR [INFO] [stderr] HOST_AR = None [INFO] [stderr] cargo:rerun-if-env-changed=AR [INFO] [stderr] AR = None [INFO] [stderr] cargo:rerun-if-env-changed=ARFLAGS_x86_64-unknown-linux-gnu [INFO] [stderr] ARFLAGS_x86_64-unknown-linux-gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=ARFLAGS_x86_64_unknown_linux_gnu [INFO] [stderr] ARFLAGS_x86_64_unknown_linux_gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=HOST_ARFLAGS [INFO] [stderr] HOST_ARFLAGS = None [INFO] [stderr] cargo:rerun-if-env-changed=ARFLAGS [INFO] [stderr] ARFLAGS = None [INFO] [stderr] cargo:rustc-link-lib=static=ggml-vulkan [INFO] [stderr] cargo:rustc-link-search=native=/opt/rustwide/target/debug/build/llama_cpp_sys-5b531407e9d1416b/out [INFO] [stderr] cargo:rerun-if-env-changed=CXXSTDLIB_x86_64-unknown-linux-gnu [INFO] [stderr] CXXSTDLIB_x86_64-unknown-linux-gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=CXXSTDLIB_x86_64_unknown_linux_gnu [INFO] [stderr] CXXSTDLIB_x86_64_unknown_linux_gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=HOST_CXXSTDLIB [INFO] [stderr] HOST_CXXSTDLIB = None [INFO] [stderr] cargo:rerun-if-env-changed=CXXSTDLIB [INFO] [stderr] CXXSTDLIB = None [INFO] [stderr] cargo:rustc-link-lib=stdc++ [INFO] [stderr] Compiling GGML.. [INFO] [stderr] TARGET = Some("x86_64-unknown-linux-gnu") [INFO] [stderr] OPT_LEVEL = Some("0") [INFO] [stderr] HOST = Some("x86_64-unknown-linux-gnu") [INFO] [stderr] cargo:rerun-if-env-changed=CC_x86_64-unknown-linux-gnu [INFO] [stderr] CC_x86_64-unknown-linux-gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=CC_x86_64_unknown_linux_gnu [INFO] [stderr] CC_x86_64_unknown_linux_gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=HOST_CC [INFO] [stderr] HOST_CC = None [INFO] [stderr] cargo:rerun-if-env-changed=CC [INFO] [stderr] CC = None [INFO] [stderr] cargo:rerun-if-env-changed=CC_ENABLE_DEBUG_OUTPUT [INFO] [stderr] cargo:warning=Compiler version doesn't include clang or GCC: "cc" "--version" [INFO] [stderr] cargo:rerun-if-env-changed=CRATE_CC_NO_DEFAULTS [INFO] [stderr] CRATE_CC_NO_DEFAULTS = None [INFO] [stderr] DEBUG = Some("true") [INFO] [stderr] cargo:rerun-if-env-changed=CFLAGS_x86_64-unknown-linux-gnu [INFO] [stderr] CFLAGS_x86_64-unknown-linux-gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=CFLAGS_x86_64_unknown_linux_gnu [INFO] [stderr] CFLAGS_x86_64_unknown_linux_gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=HOST_CFLAGS [INFO] [stderr] HOST_CFLAGS = None [INFO] [stderr] cargo:rerun-if-env-changed=CFLAGS [INFO] [stderr] CFLAGS = None [INFO] [stderr] cargo:warning=./thirdparty/llama.cpp/ggml.c:150: warning: "GGML_DEBUG" redefined [INFO] [stderr] cargo:warning= 150 | #define GGML_DEBUG 0 [INFO] [stderr] cargo:warning= | [INFO] [stderr] cargo:warning=: note: this is the location of the previous definition [INFO] [stderr] cargo:warning=./thirdparty/llama.cpp/ggml-alloc.c:448:13: warning: 'ggml_gallocr_set_node_offset' defined but not used [-Wunused-function] [INFO] [stderr] cargo:warning= 448 | static void ggml_gallocr_set_node_offset(ggml_gallocr_t galloc, struct ggml_tensor * node, int buffer_id, size_t offset) { [INFO] [stderr] cargo:warning= | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~ [INFO] [stderr] cargo:warning=./thirdparty/llama.cpp/ggml-backend.c:1067:13: warning: 'ggml_backend_sched_print_assignments' defined but not used [-Wunused-function] [INFO] [stderr] cargo:warning= 1067 | static void ggml_backend_sched_print_assignments(ggml_backend_sched_t sched, struct ggml_cgraph * graph) { [INFO] [stderr] cargo:warning= | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ [INFO] [stderr] cargo:warning=./thirdparty/llama.cpp/ggml-quants.c:1389:14: warning: 'make_qkx1_quants' defined but not used [-Wunused-function] [INFO] [stderr] cargo:warning= 1389 | static float make_qkx1_quants(int n, int nmax, const float * restrict x, uint8_t * restrict L, float * restrict the_min, [INFO] [stderr] cargo:warning= | ^~~~~~~~~~~~~~~~ [INFO] [stderr] cargo:warning=./thirdparty/llama.cpp/ggml.c:18561:13: warning: 'ggml_opt_get_grad' defined but not used [-Wunused-function] [INFO] [stderr] cargo:warning=18561 | static void ggml_opt_get_grad(int np, struct ggml_tensor * const ps[], float * g) { [INFO] [stderr] cargo:warning= | ^~~~~~~~~~~~~~~~~ [INFO] [stderr] exit status: 0 [INFO] [stderr] exit status: 0 [INFO] [stderr] exit status: 0 [INFO] [stderr] exit status: 0 [INFO] [stderr] cargo:rerun-if-env-changed=AR_x86_64-unknown-linux-gnu [INFO] [stderr] AR_x86_64-unknown-linux-gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=AR_x86_64_unknown_linux_gnu [INFO] [stderr] AR_x86_64_unknown_linux_gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=HOST_AR [INFO] [stderr] HOST_AR = None [INFO] [stderr] cargo:rerun-if-env-changed=AR [INFO] [stderr] AR = None [INFO] [stderr] cargo:rerun-if-env-changed=ARFLAGS_x86_64-unknown-linux-gnu [INFO] [stderr] ARFLAGS_x86_64-unknown-linux-gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=ARFLAGS_x86_64_unknown_linux_gnu [INFO] [stderr] ARFLAGS_x86_64_unknown_linux_gnu = None [INFO] [stderr] cargo:rerun-if-env-changed=HOST_ARFLAGS [INFO] [stderr] HOST_ARFLAGS = None [INFO] [stderr] cargo:rerun-if-env-changed=ARFLAGS [INFO] [stderr] ARFLAGS = None [INFO] [stderr] cargo:rustc-link-lib=static=ggml [INFO] [stderr] cargo:rustc-link-search=native=/opt/rustwide/target/debug/build/llama_cpp_sys-5b531407e9d1416b/out [INFO] [stderr] Compiling Llama.cpp.. [INFO] [stderr] cargo:warning=./thirdparty/llama.cpp/llama.cpp: In function 'ggml_tensor* llm_build_kqv(ggml_context*, const llama_model&, const llama_hparams&, const llama_kv_cache&, ggml_cgraph*, ggml_tensor*, ggml_tensor*, ggml_tensor*, ggml_tensor*, ggml_tensor*, int64_t, int32_t, int32_t, float, const llm_build_cb&, int)': [INFO] [stderr] cargo:warning=./thirdparty/llama.cpp/llama.cpp:4887:108: note: '#pragma message: TODO: ALiBi support in ggml_soft_max_ext is not implemented for Vulkan, Kompute, and SYCL' [INFO] [stderr] cargo:warning= 4887 | #pragma message("TODO: ALiBi support in ggml_soft_max_ext is not implemented for Vulkan, Kompute, and SYCL") [INFO] [stderr] cargo:warning= | ^ [INFO] [stderr] cargo:warning=./thirdparty/llama.cpp/llama.cpp:4888:87: note: '#pragma message: Falling back to ggml_alibi(). Will become an error in Mar 2024' [INFO] [stderr] cargo:warning= 4888 | #pragma message(" Falling back to ggml_alibi(). Will become an error in Mar 2024") [INFO] [stderr] cargo:warning= | ^ [INFO] [stderr] cargo:warning=./thirdparty/llama.cpp/llama.cpp:4889:73: note: '#pragma message: ref: https://github.com/ggerganov/llama.cpp/pull/5488' [INFO] [stderr] cargo:warning= 4889 | #pragma message("ref: https://github.com/ggerganov/llama.cpp/pull/5488") [INFO] [stderr] cargo:warning= | ^ [INFO] [stderr] cargo:warning=./thirdparty/llama.cpp/llama.cpp:4894:24: warning: 'ggml_tensor* ggml_alibi(ggml_context*, ggml_tensor*, int, int, float)' is deprecated: use ggml_soft_max_ext instead (will be removed in Mar 2024) [-Wdeprecated-declarations] [INFO] [stderr] cargo:warning= 4894 | kq = ggml_alibi(ctx, kq, /*n_past*/ 0, n_head, hparams.f_max_alibi_bias); [INFO] [stderr] cargo:warning= | ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ [INFO] [stderr] cargo:warning=In file included from ./thirdparty/llama.cpp/llama.h:4, [INFO] [stderr] cargo:warning= from ./thirdparty/llama.cpp/llama.cpp:2: [INFO] [stderr] cargo:warning=./thirdparty/llama.cpp/ggml.h:1502:51: note: declared here [INFO] [stderr] cargo:warning= 1502 | GGML_DEPRECATED(GGML_API struct ggml_tensor * ggml_alibi( [INFO] [stderr] cargo:warning= | ^~~~~~~~~~ [INFO] [stderr] cargo:warning=./thirdparty/llama.cpp/ggml.h:202:41: note: in definition of macro 'GGML_DEPRECATED' [INFO] [stderr] cargo:warning= 202 | # define GGML_DEPRECATED(func, hint) func __attribute__((deprecated(hint))) [INFO] [stderr] cargo:warning= | ^~~~ [INFO] [stderr] cargo:warning=./thirdparty/llama.cpp/llama.cpp: At global scope: [INFO] [stderr] cargo:warning=./thirdparty/llama.cpp/llama.cpp:12940:1: fatal error: error writing to /tmp/cc3lMeL6.s: No space left on device [INFO] [stderr] cargo:warning=12940 | } [INFO] [stderr] cargo:warning= | ^ [INFO] [stderr] cargo:warning=compilation terminated. [INFO] [stderr] [INFO] [stderr] --- stderr [INFO] [stderr] error: 'rustfmt' is not installed for the custom toolchain 'cccf075eba88363269e8589ebb8d40874cc542d8'. [INFO] [stderr] note: this is a custom toolchain, which cannot use `rustup component add` [INFO] [stderr] help: if you built this toolchain from source, and used `rustup toolchain link`, then you may be able to build the component with `x.py` [INFO] [stderr] Failed to run rustfmt: Internal rustfmt error (non-fatal, continuing) [INFO] [stderr] [INFO] [stderr] [INFO] [stderr] error occurred: Command "c++" "-O0" "-ffunction-sections" "-fdata-sections" "-fPIC" "-gdwarf-4" "-fno-omit-frame-pointer" "-m64" "-static" "-std=c++14" "-I" "./thirdparty/llama.cpp" "-Wall" "-Wextra" "-fPIC" "-pthread" "-Wall" "-Wdeprecated-declarations" "-Wextra" "-Wpedantic" "-Wcast-qual" "-Wno-unused-function" "-Wno-multichar" "-march=native" "-mfma" "-mf16c" "-mavx2" "-mavx" "-DGGML_DEBUG=100" "-D_GLIBCXX_ASSERTIONS" "-D_XOPEN_SOURCE=600" "-D_GNU_SOURCE" "-DGGML_VULKAN_DEBUG" "-DGGML_VULKAN_VALIDATE" "-DGGML_USE_VULKAN" "-o" "/opt/rustwide/target/debug/build/llama_cpp_sys-5b531407e9d1416b/out/50cbb087c3899f86-llama.o" "-c" "./thirdparty/llama.cpp/llama.cpp" with args "c++" did not execute successfully (status code exit status: 1). [INFO] [stderr] [INFO] [stderr] [INFO] running `Command { std: "docker" "inspect" "fe0ce4980f06865ca7ca39d8f48d47b78b507e319e6ee6007a4991f0440c67cc", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "fe0ce4980f06865ca7ca39d8f48d47b78b507e319e6ee6007a4991f0440c67cc", kill_on_drop: false }` [INFO] [stdout] fe0ce4980f06865ca7ca39d8f48d47b78b507e319e6ee6007a4991f0440c67cc