From 66330c5eaa1228dc20e8586f7109035209213b7e Mon Sep 17 00:00:00 2001 From: Vitaly Chikunov Date: Sat, 10 May 2025 02:13:53 +0300 Subject: [PATCH] 1:5332-alt1 - Update to b5332 (2025-05-09), with vision support in llama-server. - Enable Vulkan backend (for GPU) in llama.cpp-vulkan package. --- .gear/llama.cpp.spec | 6 +++--- .gear/tags/list | 2 +- 2 files changed, 4 insertions(+), 4 deletions(-) diff --git a/.gear/llama.cpp.spec b/.gear/llama.cpp.spec index 307a71e11..b3416aa2b 100644 --- a/.gear/llama.cpp.spec +++ b/.gear/llama.cpp.spec @@ -11,7 +11,7 @@ %def_with vulkan Name: llama.cpp -Version: 5318 +Version: 5332 Release: alt1 Epoch: 1 Summary: LLM inference in C/C++ @@ -251,8 +251,8 @@ llama-cli -m %_datadir/tinyllamas/stories260K.gguf -p "Once upon a time" -s 55 - %endif %changelog -* Fri May 09 2025 Vitaly Chikunov 1:5318-alt1 -- Update to b5318 (2025-05-08). +* Sat May 10 2025 Vitaly Chikunov 1:5332-alt1 +- Update to b5332 (2025-05-09), with vision support in llama-server. - Enable Vulkan backend (for GPU) in llama.cpp-vulkan package. * Mon Mar 10 2025 Vitaly Chikunov 1:4855-alt1 diff --git a/.gear/tags/list b/.gear/tags/list index 1eff7fd13..1505f6a07 100644 --- a/.gear/tags/list +++ b/.gear/tags/list @@ -1,2 +1,2 @@ 4565194ed7c32d1d2efa32ceab4d3c6cae006306 submodule/kompute-0 -15e03282bb432631193464100c2237a3b6bcfe4c b5318 +7c28a74e0783f4bb74a246fb9f19bf212139e365 b5332