diff --git a/.SRCINFO b/.SRCINFO index 339634ae66e7d..49d25a8bfdc56 100644 --- a/.SRCINFO +++ b/.SRCINFO @@ -1,6 +1,6 @@ pkgbase = llama.cpp-vulkan pkgdesc = Port of Facebook's LLaMA model in C/C++ (with Vulkan GPU optimizations) - pkgver = b5723 + pkgver = b5726 pkgrel = 1 url = https://github.com/ggerganov/llama.cpp arch = x86_64 @@ -21,10 +21,10 @@ pkgbase = llama.cpp-vulkan conflicts = llama.cpp options = lto options = !debug - source = git+https://github.com/ggerganov/llama.cpp#tag=b5723 + source = git+https://github.com/ggerganov/llama.cpp#tag=b5726 source = llama.cpp.conf source = llama.cpp.service - sha256sums = f7d5b052e3c10a94741f331dcecdfbebb0b3a4375871a86a437f74855f5fe830 + sha256sums = fe7abfa6d694a1e1fdbc390249aa41c64ffb5f97b2f2f2fa2d6379da95eb14ab sha256sums = 53fa70cfe40cb8a3ca432590e4f76561df0f129a31b121c9b4b34af0da7c4d87 sha256sums = 0377d08a07bda056785981d3352ccd2dbc0387c4836f91fb73e6b790d836620d diff --git a/PKGBUILD b/PKGBUILD index 6edea46f61f91..6b2ab880b0b02 100644 --- a/PKGBUILD +++ b/PKGBUILD @@ -2,7 +2,7 @@ pkgname=llama.cpp-vulkan _pkgname=${pkgname%%-vulkan} -pkgver=b5723 +pkgver=b5726 pkgrel=1 pkgdesc="Port of Facebook's LLaMA model in C/C++ (with Vulkan GPU optimizations)" arch=(x86_64 armv7h aarch64) @@ -30,7 +30,7 @@ source=( llama.cpp.conf llama.cpp.service ) -sha256sums=('f7d5b052e3c10a94741f331dcecdfbebb0b3a4375871a86a437f74855f5fe830' +sha256sums=('fe7abfa6d694a1e1fdbc390249aa41c64ffb5f97b2f2f2fa2d6379da95eb14ab' '53fa70cfe40cb8a3ca432590e4f76561df0f129a31b121c9b4b34af0da7c4d87' '0377d08a07bda056785981d3352ccd2dbc0387c4836f91fb73e6b790d836620d')