1:5753-alt1
- Update to b5753 (2025-06-24). - Install an experimental rpc backend and server. The rpc code is a proof-of-concept, fragile, and insecure.
This commit is contained in:
parent
7207d978c8
commit
ce184d9ae8
2 changed files with 13 additions and 6 deletions
|
|
@ -11,7 +11,7 @@
|
|||
%def_with vulkan
|
||||
|
||||
Name: llama.cpp
|
||||
Version: 5332
|
||||
Version: 5753
|
||||
Release: alt1
|
||||
Epoch: 1
|
||||
Summary: LLM inference in C/C++
|
||||
|
|
@ -147,11 +147,11 @@ cat <<-EOF >> cmake/build-info.cmake
|
|||
set(BUILD_COMMIT "${commit::8} [%release]")
|
||||
EOF
|
||||
sed -i '/POSITION_INDEPENDENT_CODE/s/PROPERTIES/& SOVERSION 0.0.%version/' ggml/src/CMakeLists.txt src/CMakeLists.txt
|
||||
sed -i 's/POSITION_INDEPENDENT_CODE/SOVERSION 0.0.%version &/' ggml/cmake/ggml-config.cmake.in
|
||||
sed -i 's/POSITION_INDEPENDENT_CODE/SOVERSION 0.0.%version &/' ggml/cmake/ggml-config.cmake.in tools/mtmd/CMakeLists.txt
|
||||
# We do not have Internet access (issues/13371).
|
||||
sed -i 's/common_has_curl()/0/' tests/test-arg-parser.cpp
|
||||
# Libs with unclear purpose.
|
||||
sed -i s/BUILD_SHARED_LIBS/0/ tools/mtmd/CMakeLists.txt
|
||||
# This test requires GPU.
|
||||
sed /test-thread-safety/d -i tests/CMakeLists.txt
|
||||
|
||||
%build
|
||||
# Unless -DCMAKE_SKIP_BUILD_RPATH=yes CMake fails to strip build time RPATH
|
||||
|
|
@ -217,14 +217,17 @@ llama-cli -m %_datadir/tinyllamas/stories260K.gguf -p "Once upon a time" -s 55 -
|
|||
%_libdir/libllama.so.0.0.%version
|
||||
%_libdir/libggml.so.0.0.%version
|
||||
%_libdir/libggml-base.so.0.0.%version
|
||||
%_libdir/libmtmd.so.0.0.%version
|
||||
|
||||
%files -n libllama-devel
|
||||
%_libdir/libllama.so
|
||||
%_libdir/libggml.so
|
||||
%_libdir/libggml-base.so
|
||||
%_libdir/libmtmd.so
|
||||
%_includedir/llama*.h
|
||||
%_includedir/gguf.h
|
||||
%_includedir/ggml*.h
|
||||
%_includedir/mtmd*.h
|
||||
%_cmakedir/ggml
|
||||
%_cmakedir/llama
|
||||
%_pkgconfigdir/llama.pc
|
||||
|
|
@ -248,12 +251,16 @@ llama-cli -m %_datadir/tinyllamas/stories260K.gguf -p "Once upon a time" -s 55 -
|
|||
|
||||
%if_with vulkan
|
||||
%files vulkan
|
||||
%_bindir/vulkan-shaders-gen
|
||||
%dir %_libexecdir/llama
|
||||
%_libexecdir/llama/libggml-vulkan.so
|
||||
%endif
|
||||
|
||||
%changelog
|
||||
* Wed Jun 25 2025 Vitaly Chikunov <vt@altlinux.org> 1:5753-alt1
|
||||
- Update to b5753 (2025-06-24).
|
||||
- Install an experimental rpc backend and server. The rpc code is a
|
||||
proof-of-concept, fragile, and insecure.
|
||||
|
||||
* Sat May 10 2025 Vitaly Chikunov <vt@altlinux.org> 1:5332-alt1
|
||||
- Update to b5332 (2025-05-09), with vision support in llama-server.
|
||||
- Enable Vulkan backend (for GPU) in llama.cpp-vulkan package.
|
||||
|
|
|
|||
|
|
@ -1,2 +1,2 @@
|
|||
4565194ed7c32d1d2efa32ceab4d3c6cae006306 submodule/kompute-0
|
||||
7c28a74e0783f4bb74a246fb9f19bf212139e365 b5332
|
||||
73e53dc834c0a2336cd104473af6897197b96277 b5753
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue