From 0a9a879c97f8513e9269786ad16cbf9e3fbcde07 Mon Sep 17 00:00:00 2001 From: Orion Date: Fri, 26 Dec 2025 00:23:10 +0000 Subject: [PATCH] :bookmark: Bump ik-llama.cpp-cuda version to r4079.03ed5f70-1 --- .SRCINFO | 2 +- PKGBUILD | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/.SRCINFO b/.SRCINFO index 4f2fba9ebd94e..523c32cf1457e 100644 --- a/.SRCINFO +++ b/.SRCINFO @@ -1,6 +1,6 @@ pkgbase = ik-llama.cpp-cuda pkgdesc = llama.cpp fork with additional SOTA quants and improved performance (CUDA Backend) - pkgver = r4077.fbb67fa2 + pkgver = r4079.03ed5f70 pkgrel = 1 url = https://github.com/ikawrakow/ik_llama.cpp arch = x86_64 diff --git a/PKGBUILD b/PKGBUILD index 824e1fac3f9b2..79eae62f14765 100644 --- a/PKGBUILD +++ b/PKGBUILD @@ -2,7 +2,7 @@ pkgname=ik-llama.cpp-cuda _pkgname=ik_llama.cpp -pkgver=r4077.fbb67fa2 +pkgver=r4079.03ed5f70 pkgrel=1 pkgdesc="llama.cpp fork with additional SOTA quants and improved performance (CUDA Backend)" arch=(x86_64 armv7h aarch64)