diff options
Diffstat (limited to 'misc/llama-cpp')
-rw-r--r-- | misc/llama-cpp/Makefile | 3 | ||||
-rw-r--r-- | misc/llama-cpp/distinfo | 6 | ||||
-rw-r--r-- | misc/llama-cpp/pkg-plist | 9 |
3 files changed, 10 insertions, 8 deletions
diff --git a/misc/llama-cpp/Makefile b/misc/llama-cpp/Makefile index e880ababbdc1..463389d10d70 100644 --- a/misc/llama-cpp/Makefile +++ b/misc/llama-cpp/Makefile @@ -1,6 +1,6 @@ PORTNAME= llama-cpp DISTVERSIONPREFIX= b -DISTVERSION= 5371 +DISTVERSION= 6374 CATEGORIES= misc # machine-learning MAINTAINER= yuri@FreeBSD.org @@ -44,7 +44,6 @@ CURL_LIB_DEPENDS= libcurl.so:ftp/curl EXAMPLES_CMAKE_BOOL= LLAMA_BUILD_EXAMPLES -VULKAN_DESC= Vulkan GPU offload support VULKAN_CMAKE_BOOL= GGML_VULKAN VULKAN_BUILD_DEPENDS= glslc:graphics/shaderc \ vulkan-headers>0:graphics/vulkan-headers diff --git a/misc/llama-cpp/distinfo b/misc/llama-cpp/distinfo index 065e1989894b..ac00c83b7dba 100644 --- a/misc/llama-cpp/distinfo +++ b/misc/llama-cpp/distinfo @@ -1,5 +1,5 @@ -TIMESTAMP = 1747201270 -SHA256 (ggerganov-llama.cpp-b5371_GH0.tar.gz) = cce50220507565b78423fc45a1c534dc088289ab898517a379fdbf733ffd72bf -SIZE (ggerganov-llama.cpp-b5371_GH0.tar.gz) = 21147325 +TIMESTAMP = 1756957718 +SHA256 (ggerganov-llama.cpp-b6374_GH0.tar.gz) = fe4392f8331a13fcd6ccf5fa4a7e9a79d58958394ef62bf8df6d214208619ed5 +SIZE (ggerganov-llama.cpp-b6374_GH0.tar.gz) = 25632053 SHA256 (nomic-ai-kompute-4565194_GH0.tar.gz) = 95b52d2f0514c5201c7838348a9c3c9e60902ea3c6c9aa862193a212150b2bfc SIZE (nomic-ai-kompute-4565194_GH0.tar.gz) = 13540496 diff --git a/misc/llama-cpp/pkg-plist b/misc/llama-cpp/pkg-plist index 5b36c5809aaa..9c53359aabc9 100644 --- a/misc/llama-cpp/pkg-plist +++ b/misc/llama-cpp/pkg-plist @@ -5,6 +5,7 @@ bin/convert_hf_to_gguf.py %%EXAMPLES%%bin/llama-cli %%EXAMPLES%%bin/llama-convert-llama2c-to-ggml %%EXAMPLES%%bin/llama-cvector-generator +%%EXAMPLES%%bin/llama-diffusion-cli %%EXAMPLES%%bin/llama-embedding %%EXAMPLES%%bin/llama-eval-callback %%EXAMPLES%%bin/llama-export-lora @@ -15,6 +16,7 @@ bin/convert_hf_to_gguf.py %%EXAMPLES%%bin/llama-gguf-split %%EXAMPLES%%bin/llama-gritlm %%EXAMPLES%%bin/llama-imatrix +%%EXAMPLES%%bin/llama-logits %%EXAMPLES%%bin/llama-lookahead %%EXAMPLES%%bin/llama-lookup %%EXAMPLES%%bin/llama-lookup-create @@ -35,7 +37,6 @@ bin/convert_hf_to_gguf.py %%EXAMPLES%%bin/llama-speculative-simple %%EXAMPLES%%bin/llama-tokenize %%EXAMPLES%%bin/llama-tts -%%VULKAN%%bin/vulkan-shaders-gen include/ggml-alloc.h include/ggml-backend.h include/ggml-blas.h @@ -43,16 +44,18 @@ include/ggml-cann.h include/ggml-cpp.h include/ggml-cpu.h include/ggml-cuda.h -include/ggml-kompute.h include/ggml-metal.h include/ggml-opt.h include/ggml-rpc.h include/ggml-sycl.h include/ggml-vulkan.h +include/ggml-webgpu.h include/ggml.h include/gguf.h include/llama-cpp.h include/llama.h +include/mtmd-helper.h +include/mtmd.h lib/cmake/ggml/ggml-config.cmake lib/cmake/ggml/ggml-version.cmake lib/cmake/llama/llama-config.cmake @@ -62,5 +65,5 @@ lib/libggml-cpu.so %%VULKAN%%lib/libggml-vulkan.so lib/libggml.so lib/libllama.so -lib/libmtmd_shared.so +lib/libmtmd.so libdata/pkgconfig/llama.pc |