diff options
Diffstat (limited to 'misc/llama-cpp')
| -rw-r--r-- | misc/llama-cpp/Makefile | 4 | ||||
| -rw-r--r-- | misc/llama-cpp/distinfo | 6 | ||||
| -rw-r--r-- | misc/llama-cpp/files/llama-server.in | 4 | ||||
| -rw-r--r-- | misc/llama-cpp/files/patch-ggml_src_CMakeLists.txt | 20 | ||||
| -rw-r--r-- | misc/llama-cpp/pkg-plist | 27 |
5 files changed, 29 insertions, 32 deletions
diff --git a/misc/llama-cpp/Makefile b/misc/llama-cpp/Makefile index e880ababbdc1..12023ccb8981 100644 --- a/misc/llama-cpp/Makefile +++ b/misc/llama-cpp/Makefile @@ -1,6 +1,6 @@ PORTNAME= llama-cpp DISTVERSIONPREFIX= b -DISTVERSION= 5371 +DISTVERSION= 7709 CATEGORIES= misc # machine-learning MAINTAINER= yuri@FreeBSD.org @@ -24,6 +24,7 @@ GH_TUPLE= nomic-ai:kompute:4565194:kompute/kompute SHEBANG_GLOB= *.py CMAKE_ON= BUILD_SHARED_LIBS + #LLAMA_USE_SYSTEM_GGML CMAKE_OFF= GGML_NATIVE \ FREEBSD_ALLOW_ADVANCED_CPU_FEATURES \ LLAMA_BUILD_TESTS @@ -44,7 +45,6 @@ CURL_LIB_DEPENDS= libcurl.so:ftp/curl EXAMPLES_CMAKE_BOOL= LLAMA_BUILD_EXAMPLES -VULKAN_DESC= Vulkan GPU offload support VULKAN_CMAKE_BOOL= GGML_VULKAN VULKAN_BUILD_DEPENDS= glslc:graphics/shaderc \ vulkan-headers>0:graphics/vulkan-headers diff --git a/misc/llama-cpp/distinfo b/misc/llama-cpp/distinfo index 065e1989894b..13090992cbe0 100644 --- a/misc/llama-cpp/distinfo +++ b/misc/llama-cpp/distinfo @@ -1,5 +1,5 @@ -TIMESTAMP = 1747201270 -SHA256 (ggerganov-llama.cpp-b5371_GH0.tar.gz) = cce50220507565b78423fc45a1c534dc088289ab898517a379fdbf733ffd72bf -SIZE (ggerganov-llama.cpp-b5371_GH0.tar.gz) = 21147325 +TIMESTAMP = 1768206387 +SHA256 (ggerganov-llama.cpp-b7709_GH0.tar.gz) = 8aa5d02ec90c70fa496cc878ef3962733e74184a8e43f191db5471288f5cf911 +SIZE (ggerganov-llama.cpp-b7709_GH0.tar.gz) = 28712220 SHA256 (nomic-ai-kompute-4565194_GH0.tar.gz) = 95b52d2f0514c5201c7838348a9c3c9e60902ea3c6c9aa862193a212150b2bfc SIZE (nomic-ai-kompute-4565194_GH0.tar.gz) = 13540496 diff --git a/misc/llama-cpp/files/llama-server.in b/misc/llama-cpp/files/llama-server.in index f958b774a4fc..6eec15953978 100644 --- a/misc/llama-cpp/files/llama-server.in +++ b/misc/llama-cpp/files/llama-server.in @@ -55,8 +55,8 @@ llama_server_precmd() fi # ensure that the log file exists and has right permissions - touch ${llama_server_log} - chown ${llama_server_user} ${llama_server_log} + touch ${llama_server_log} ${pidfile} + chown ${llama_server_user} ${llama_server_log} ${pidfile} chmod 640 ${llama_server_log} } diff --git a/misc/llama-cpp/files/patch-ggml_src_CMakeLists.txt b/misc/llama-cpp/files/patch-ggml_src_CMakeLists.txt deleted file mode 100644 index 8135824e1b8b..000000000000 --- a/misc/llama-cpp/files/patch-ggml_src_CMakeLists.txt +++ /dev/null @@ -1,20 +0,0 @@ -- workaround for https://github.com/ggerganov/llama.cpp/issues/11095 - ---- ggml/src/CMakeLists.txt.orig 2025-01-06 00:37:35 UTC -+++ ggml/src/CMakeLists.txt -@@ -152,15 +152,6 @@ endif() - # posix_memalign came in POSIX.1-2001 / SUSv3 - # M_PI is an XSI extension since POSIX.1-2001 / SUSv3, came in XPG1 (1985) - --# Somehow in OpenBSD whenever POSIX conformance is specified --# some string functions rely on locale_t availability, --# which was introduced in POSIX.1-2008, forcing us to go higher --if (CMAKE_SYSTEM_NAME MATCHES "OpenBSD") -- add_compile_definitions(_XOPEN_SOURCE=700) --else() -- add_compile_definitions(_XOPEN_SOURCE=600) --endif() -- - # Data types, macros and functions related to controlling CPU affinity and - # some memory allocation are available on Linux through GNU extensions in libc - if (CMAKE_SYSTEM_NAME MATCHES "Linux" OR CMAKE_SYSTEM_NAME MATCHES "Android") diff --git a/misc/llama-cpp/pkg-plist b/misc/llama-cpp/pkg-plist index 5b36c5809aaa..84f8c70f2bbd 100644 --- a/misc/llama-cpp/pkg-plist +++ b/misc/llama-cpp/pkg-plist @@ -3,17 +3,21 @@ bin/convert_hf_to_gguf.py %%EXAMPLES%%bin/llama-batched-bench %%EXAMPLES%%bin/llama-bench %%EXAMPLES%%bin/llama-cli +%%EXAMPLES%%bin/llama-completion %%EXAMPLES%%bin/llama-convert-llama2c-to-ggml %%EXAMPLES%%bin/llama-cvector-generator +%%EXAMPLES%%bin/llama-debug +%%EXAMPLES%%bin/llama-diffusion-cli %%EXAMPLES%%bin/llama-embedding %%EXAMPLES%%bin/llama-eval-callback %%EXAMPLES%%bin/llama-export-lora %%EXAMPLES%%bin/llama-finetune +%%EXAMPLES%%bin/llama-fit-params %%EXAMPLES%%bin/llama-gen-docs %%EXAMPLES%%bin/llama-gguf %%EXAMPLES%%bin/llama-gguf-hash %%EXAMPLES%%bin/llama-gguf-split -%%EXAMPLES%%bin/llama-gritlm +%%EXAMPLES%%bin/llama-idle %%EXAMPLES%%bin/llama-imatrix %%EXAMPLES%%bin/llama-lookahead %%EXAMPLES%%bin/llama-lookup @@ -26,7 +30,6 @@ bin/convert_hf_to_gguf.py %%EXAMPLES%%bin/llama-perplexity %%EXAMPLES%%bin/llama-quantize %%EXAMPLES%%bin/llama-retrieval -%%EXAMPLES%%bin/llama-run %%EXAMPLES%%bin/llama-save-load-state %%EXAMPLES%%bin/llama-server %%EXAMPLES%%bin/llama-simple @@ -35,7 +38,6 @@ bin/convert_hf_to_gguf.py %%EXAMPLES%%bin/llama-speculative-simple %%EXAMPLES%%bin/llama-tokenize %%EXAMPLES%%bin/llama-tts -%%VULKAN%%bin/vulkan-shaders-gen include/ggml-alloc.h include/ggml-backend.h include/ggml-blas.h @@ -43,24 +45,39 @@ include/ggml-cann.h include/ggml-cpp.h include/ggml-cpu.h include/ggml-cuda.h -include/ggml-kompute.h include/ggml-metal.h include/ggml-opt.h include/ggml-rpc.h include/ggml-sycl.h include/ggml-vulkan.h +include/ggml-webgpu.h +include/ggml-zendnn.h include/ggml.h include/gguf.h include/llama-cpp.h include/llama.h +include/mtmd-helper.h +include/mtmd.h lib/cmake/ggml/ggml-config.cmake lib/cmake/ggml/ggml-version.cmake lib/cmake/llama/llama-config.cmake lib/cmake/llama/llama-version.cmake lib/libggml-base.so +lib/libggml-base.so.0 +lib/libggml-base.so.0.9.5 lib/libggml-cpu.so +lib/libggml-cpu.so.0 +lib/libggml-cpu.so.0.9.5 %%VULKAN%%lib/libggml-vulkan.so +%%VULKAN%%lib/libggml-vulkan.so.0 +%%VULKAN%%lib/libggml-vulkan.so.0.9.5 lib/libggml.so +lib/libggml.so.0 +lib/libggml.so.0.9.5 lib/libllama.so -lib/libmtmd_shared.so +lib/libllama.so.0 +lib/libllama.so.0.0.7709 +lib/libmtmd.so +lib/libmtmd.so.0 +lib/libmtmd.so.0.0.7709 libdata/pkgconfig/llama.pc |
