diff --git a/.gear/llama.cpp.spec b/.gear/llama.cpp.spec index da0c07826..bd89cf31d 100644 --- a/.gear/llama.cpp.spec +++ b/.gear/llama.cpp.spec @@ -11,7 +11,7 @@ %def_with vulkan Name: llama.cpp -Version: 5753 +Version: 6121 Release: alt1 Epoch: 1 Summary: LLM inference in C/C++ @@ -160,6 +160,7 @@ export NVCC_PREPEND_FLAGS=-ccbin=g++-12 -DLLAMA_BUILD_TESTS=ON \ -DLLAMA_CURL=ON \ -DGGML_BACKEND_DL=ON \ + -DGGML_BACKEND_DIR=%_libexecdir/llama \ -DGGML_CPU=ON \ -DGGML_RPC=ON \ %ifarch x86_64 @@ -256,6 +257,9 @@ llama-cli -m %_datadir/tinyllamas/stories260K.gguf -p "Once upon a time" -s 55 - %endif %changelog +* Sat Aug 09 2025 Vitaly Chikunov 1:6121-alt1 +- Update to b6121 (2025-08-08). + * Wed Jun 25 2025 Vitaly Chikunov 1:5753-alt1 - Update to b5753 (2025-06-24). - Install an experimental rpc backend and server. The rpc code is a diff --git a/.gear/tags/list b/.gear/tags/list index d58fb2545..6fde054ce 100644 --- a/.gear/tags/list +++ b/.gear/tags/list @@ -1 +1 @@ -73e53dc834c0a2336cd104473af6897197b96277 b5753 +e54d41befcc1575f4c898c5ff4ef43970cead75f b6121