diff --git a/packages/llama-cpp/0002-do-not-install-backend-libs.patch b/packages/llama-cpp/0002-do-not-install-backend-libs.patch index ff4c1498f5..c56fc380e1 100644 --- a/packages/llama-cpp/0002-do-not-install-backend-libs.patch +++ b/packages/llama-cpp/0002-do-not-install-backend-libs.patch @@ -1,10 +1,10 @@ --- a/ggml/src/CMakeLists.txt +++ b/ggml/src/CMakeLists.txt -@@ -227,7 +227,6 @@ - set_target_properties(${backend} PROPERTIES LIBRARY_OUTPUT_DIRECTORY ${CMAKE_RUNTIME_OUTPUT_DIRECTORY}) - target_compile_definitions(${backend} PRIVATE GGML_BACKEND_DL) - add_dependencies(ggml ${backend}) -- install(TARGETS ${backend} LIBRARY DESTINATION ${CMAKE_INSTALL_BINDIR}) +@@ -237,7 +237,6 @@ + if (GGML_BACKEND_DIR) + install(TARGETS ${backend} LIBRARY DESTINATION ${GGML_BACKEND_DIR}) + else() +- install(TARGETS ${backend} LIBRARY DESTINATION ${CMAKE_INSTALL_BINDIR}) + endif() else() add_library(${backend} ${ARGN}) - target_link_libraries(ggml PUBLIC ${backend}) diff --git a/packages/llama-cpp/build.sh b/packages/llama-cpp/build.sh index 7834d58e4f..d279909f6e 100644 --- a/packages/llama-cpp/build.sh +++ b/packages/llama-cpp/build.sh @@ -2,9 +2,9 @@ TERMUX_PKG_HOMEPAGE=https://github.com/ggml-org/llama.cpp TERMUX_PKG_DESCRIPTION="LLM inference in C/C++" TERMUX_PKG_LICENSE="MIT" TERMUX_PKG_MAINTAINER=@termux -TERMUX_PKG_VERSION="0.0.0-b6084" +TERMUX_PKG_VERSION="0.0.0-b6089" TERMUX_PKG_SRCURL=https://github.com/ggml-org/llama.cpp/archive/refs/tags/${TERMUX_PKG_VERSION#*-}.tar.gz -TERMUX_PKG_SHA256=6865dac18586976dacd43247f9ad2e655a46eaca017392517e0682c9f0cd1a06 +TERMUX_PKG_SHA256=6b32557024e487545792397d7620561951012bf2a3def54f979835ab097c488b TERMUX_PKG_AUTO_UPDATE=true TERMUX_PKG_DEPENDS="libc++, libcurl" TERMUX_PKG_BUILD_DEPENDS="vulkan-headers, opencl-headers, ocl-icd"