From 6543c9b82a7f81ee4e1a4d161affb30e5ac1822f Mon Sep 17 00:00:00 2001 From: Orion Date: Mon, 1 Sep 2025 00:42:54 +0000 Subject: [PATCH] :bookmark: Bump ik-llama.cpp-cuda version to r3869.b66cecca-1 --- .SRCINFO | 2 +- PKGBUILD | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/.SRCINFO b/.SRCINFO index 42983e38b3d00..9d07c8335aa37 100644 --- a/.SRCINFO +++ b/.SRCINFO @@ -1,6 +1,6 @@ pkgbase = ik-llama.cpp-cuda pkgdesc = llama.cpp fork with additional SOTA quants and improved performance (CUDA Backend) - pkgver = r3858.af13c9a2 + pkgver = r3869.b66cecca pkgrel = 1 url = https://github.com/ikawrakow/ik_llama.cpp arch = x86_64 diff --git a/PKGBUILD b/PKGBUILD index 0340a9af2040c..dde5c9eabcebd 100644 --- a/PKGBUILD +++ b/PKGBUILD @@ -2,7 +2,7 @@ pkgname=ik-llama.cpp-cuda _pkgname=ik_llama.cpp -pkgver=r3858.af13c9a2 +pkgver=r3869.b66cecca pkgrel=1 pkgdesc="llama.cpp fork with additional SOTA quants and improved performance (CUDA Backend)" arch=(x86_64 armv7h aarch64)