summaryrefslogtreecommitdiff
path: root/misc/llama-cpp
diff options
context:
space:
mode:
Diffstat (limited to 'misc/llama-cpp')
-rw-r--r--misc/llama-cpp/Makefile4
-rw-r--r--misc/llama-cpp/distinfo6
-rw-r--r--misc/llama-cpp/files/llama-server.in4
-rw-r--r--misc/llama-cpp/files/patch-ggml_src_CMakeLists.txt20
-rw-r--r--misc/llama-cpp/pkg-plist27
5 files changed, 29 insertions, 32 deletions
diff --git a/misc/llama-cpp/Makefile b/misc/llama-cpp/Makefile
index e880ababbdc1..12023ccb8981 100644
--- a/misc/llama-cpp/Makefile
+++ b/misc/llama-cpp/Makefile
@@ -1,6 +1,6 @@
PORTNAME= llama-cpp
DISTVERSIONPREFIX= b
-DISTVERSION= 5371
+DISTVERSION= 7709
CATEGORIES= misc # machine-learning
MAINTAINER= yuri@FreeBSD.org
@@ -24,6 +24,7 @@ GH_TUPLE= nomic-ai:kompute:4565194:kompute/kompute
SHEBANG_GLOB= *.py
CMAKE_ON= BUILD_SHARED_LIBS
+ #LLAMA_USE_SYSTEM_GGML
CMAKE_OFF= GGML_NATIVE \
FREEBSD_ALLOW_ADVANCED_CPU_FEATURES \
LLAMA_BUILD_TESTS
@@ -44,7 +45,6 @@ CURL_LIB_DEPENDS= libcurl.so:ftp/curl
EXAMPLES_CMAKE_BOOL= LLAMA_BUILD_EXAMPLES
-VULKAN_DESC= Vulkan GPU offload support
VULKAN_CMAKE_BOOL= GGML_VULKAN
VULKAN_BUILD_DEPENDS= glslc:graphics/shaderc \
vulkan-headers>0:graphics/vulkan-headers
diff --git a/misc/llama-cpp/distinfo b/misc/llama-cpp/distinfo
index 065e1989894b..13090992cbe0 100644
--- a/misc/llama-cpp/distinfo
+++ b/misc/llama-cpp/distinfo
@@ -1,5 +1,5 @@
-TIMESTAMP = 1747201270
-SHA256 (ggerganov-llama.cpp-b5371_GH0.tar.gz) = cce50220507565b78423fc45a1c534dc088289ab898517a379fdbf733ffd72bf
-SIZE (ggerganov-llama.cpp-b5371_GH0.tar.gz) = 21147325
+TIMESTAMP = 1768206387
+SHA256 (ggerganov-llama.cpp-b7709_GH0.tar.gz) = 8aa5d02ec90c70fa496cc878ef3962733e74184a8e43f191db5471288f5cf911
+SIZE (ggerganov-llama.cpp-b7709_GH0.tar.gz) = 28712220
SHA256 (nomic-ai-kompute-4565194_GH0.tar.gz) = 95b52d2f0514c5201c7838348a9c3c9e60902ea3c6c9aa862193a212150b2bfc
SIZE (nomic-ai-kompute-4565194_GH0.tar.gz) = 13540496
diff --git a/misc/llama-cpp/files/llama-server.in b/misc/llama-cpp/files/llama-server.in
index f958b774a4fc..6eec15953978 100644
--- a/misc/llama-cpp/files/llama-server.in
+++ b/misc/llama-cpp/files/llama-server.in
@@ -55,8 +55,8 @@ llama_server_precmd()
fi
# ensure that the log file exists and has right permissions
- touch ${llama_server_log}
- chown ${llama_server_user} ${llama_server_log}
+ touch ${llama_server_log} ${pidfile}
+ chown ${llama_server_user} ${llama_server_log} ${pidfile}
chmod 640 ${llama_server_log}
}
diff --git a/misc/llama-cpp/files/patch-ggml_src_CMakeLists.txt b/misc/llama-cpp/files/patch-ggml_src_CMakeLists.txt
deleted file mode 100644
index 8135824e1b8b..000000000000
--- a/misc/llama-cpp/files/patch-ggml_src_CMakeLists.txt
+++ /dev/null
@@ -1,20 +0,0 @@
-- workaround for https://github.com/ggerganov/llama.cpp/issues/11095
-
---- ggml/src/CMakeLists.txt.orig 2025-01-06 00:37:35 UTC
-+++ ggml/src/CMakeLists.txt
-@@ -152,15 +152,6 @@ endif()
- # posix_memalign came in POSIX.1-2001 / SUSv3
- # M_PI is an XSI extension since POSIX.1-2001 / SUSv3, came in XPG1 (1985)
-
--# Somehow in OpenBSD whenever POSIX conformance is specified
--# some string functions rely on locale_t availability,
--# which was introduced in POSIX.1-2008, forcing us to go higher
--if (CMAKE_SYSTEM_NAME MATCHES "OpenBSD")
-- add_compile_definitions(_XOPEN_SOURCE=700)
--else()
-- add_compile_definitions(_XOPEN_SOURCE=600)
--endif()
--
- # Data types, macros and functions related to controlling CPU affinity and
- # some memory allocation are available on Linux through GNU extensions in libc
- if (CMAKE_SYSTEM_NAME MATCHES "Linux" OR CMAKE_SYSTEM_NAME MATCHES "Android")
diff --git a/misc/llama-cpp/pkg-plist b/misc/llama-cpp/pkg-plist
index 5b36c5809aaa..84f8c70f2bbd 100644
--- a/misc/llama-cpp/pkg-plist
+++ b/misc/llama-cpp/pkg-plist
@@ -3,17 +3,21 @@ bin/convert_hf_to_gguf.py
%%EXAMPLES%%bin/llama-batched-bench
%%EXAMPLES%%bin/llama-bench
%%EXAMPLES%%bin/llama-cli
+%%EXAMPLES%%bin/llama-completion
%%EXAMPLES%%bin/llama-convert-llama2c-to-ggml
%%EXAMPLES%%bin/llama-cvector-generator
+%%EXAMPLES%%bin/llama-debug
+%%EXAMPLES%%bin/llama-diffusion-cli
%%EXAMPLES%%bin/llama-embedding
%%EXAMPLES%%bin/llama-eval-callback
%%EXAMPLES%%bin/llama-export-lora
%%EXAMPLES%%bin/llama-finetune
+%%EXAMPLES%%bin/llama-fit-params
%%EXAMPLES%%bin/llama-gen-docs
%%EXAMPLES%%bin/llama-gguf
%%EXAMPLES%%bin/llama-gguf-hash
%%EXAMPLES%%bin/llama-gguf-split
-%%EXAMPLES%%bin/llama-gritlm
+%%EXAMPLES%%bin/llama-idle
%%EXAMPLES%%bin/llama-imatrix
%%EXAMPLES%%bin/llama-lookahead
%%EXAMPLES%%bin/llama-lookup
@@ -26,7 +30,6 @@ bin/convert_hf_to_gguf.py
%%EXAMPLES%%bin/llama-perplexity
%%EXAMPLES%%bin/llama-quantize
%%EXAMPLES%%bin/llama-retrieval
-%%EXAMPLES%%bin/llama-run
%%EXAMPLES%%bin/llama-save-load-state
%%EXAMPLES%%bin/llama-server
%%EXAMPLES%%bin/llama-simple
@@ -35,7 +38,6 @@ bin/convert_hf_to_gguf.py
%%EXAMPLES%%bin/llama-speculative-simple
%%EXAMPLES%%bin/llama-tokenize
%%EXAMPLES%%bin/llama-tts
-%%VULKAN%%bin/vulkan-shaders-gen
include/ggml-alloc.h
include/ggml-backend.h
include/ggml-blas.h
@@ -43,24 +45,39 @@ include/ggml-cann.h
include/ggml-cpp.h
include/ggml-cpu.h
include/ggml-cuda.h
-include/ggml-kompute.h
include/ggml-metal.h
include/ggml-opt.h
include/ggml-rpc.h
include/ggml-sycl.h
include/ggml-vulkan.h
+include/ggml-webgpu.h
+include/ggml-zendnn.h
include/ggml.h
include/gguf.h
include/llama-cpp.h
include/llama.h
+include/mtmd-helper.h
+include/mtmd.h
lib/cmake/ggml/ggml-config.cmake
lib/cmake/ggml/ggml-version.cmake
lib/cmake/llama/llama-config.cmake
lib/cmake/llama/llama-version.cmake
lib/libggml-base.so
+lib/libggml-base.so.0
+lib/libggml-base.so.0.9.5
lib/libggml-cpu.so
+lib/libggml-cpu.so.0
+lib/libggml-cpu.so.0.9.5
%%VULKAN%%lib/libggml-vulkan.so
+%%VULKAN%%lib/libggml-vulkan.so.0
+%%VULKAN%%lib/libggml-vulkan.so.0.9.5
lib/libggml.so
+lib/libggml.so.0
+lib/libggml.so.0.9.5
lib/libllama.so
-lib/libmtmd_shared.so
+lib/libllama.so.0
+lib/libllama.so.0.0.7709
+lib/libmtmd.so
+lib/libmtmd.so.0
+lib/libmtmd.so.0.0.7709
libdata/pkgconfig/llama.pc