From bd6b8f228e96a3e2681dfaa41ccf5d04f71a6768 Mon Sep 17 00:00:00 2001 From: termux-pacman-bot Date: Mon, 20 Oct 2025 00:50:19 +0000 Subject: [PATCH] bump(main/llama-cpp): 0.0.0-b6800 This commit has been automatically submitted by Github Actions. --- packages/llama-cpp/build.sh | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/packages/llama-cpp/build.sh b/packages/llama-cpp/build.sh index 5bd9b240a3..e4e3e430c1 100644 --- a/packages/llama-cpp/build.sh +++ b/packages/llama-cpp/build.sh @@ -2,9 +2,9 @@ TERMUX_PKG_HOMEPAGE=https://github.com/ggml-org/llama.cpp TERMUX_PKG_DESCRIPTION="LLM inference in C/C++" TERMUX_PKG_LICENSE="MIT" TERMUX_PKG_MAINTAINER=@termux -TERMUX_PKG_VERSION="0.0.0-b6795" +TERMUX_PKG_VERSION="0.0.0-b6800" TERMUX_PKG_SRCURL=https://github.com/ggml-org/llama.cpp/archive/refs/tags/${TERMUX_PKG_VERSION#*-}.tar.gz -TERMUX_PKG_SHA256=b037e5ecc5876e8c01c6b0a3010103b43d4883c3b1bc93fa60f09a751f256133 +TERMUX_PKG_SHA256=eaf2a570d19aa19334856589153ac6fe0405a673451c8e1dc88b0f5add5ffc4c TERMUX_PKG_AUTO_UPDATE=true TERMUX_PKG_DEPENDS="libc++, libcurl" TERMUX_PKG_BUILD_DEPENDS="vulkan-headers, opencl-headers, ocl-icd"