From ed4189066b3f13c79276a8e4bc70d200c30ac8fe Mon Sep 17 00:00:00 2001 From: Orion Date: Fri, 14 Nov 2025 00:22:31 +0000 Subject: [PATCH] :bookmark: Bump ik-llama.cpp-cuda version to r3988.be1a8cb9-1 --- .SRCINFO | 2 +- PKGBUILD | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/.SRCINFO b/.SRCINFO index fdbd8c6ff424d..da017012e4807 100644 --- a/.SRCINFO +++ b/.SRCINFO @@ -1,6 +1,6 @@ pkgbase = ik-llama.cpp-cuda pkgdesc = llama.cpp fork with additional SOTA quants and improved performance (CUDA Backend) - pkgver = r3983.8a8de91a + pkgver = r3988.be1a8cb9 pkgrel = 1 url = https://github.com/ikawrakow/ik_llama.cpp arch = x86_64 diff --git a/PKGBUILD b/PKGBUILD index 49978d17f771b..35fbdcc99a743 100644 --- a/PKGBUILD +++ b/PKGBUILD @@ -2,7 +2,7 @@ pkgname=ik-llama.cpp-cuda _pkgname=ik_llama.cpp -pkgver=r3983.8a8de91a +pkgver=r3988.be1a8cb9 pkgrel=1 pkgdesc="llama.cpp fork with additional SOTA quants and improved performance (CUDA Backend)" arch=(x86_64 armv7h aarch64)