From 542b4f899a7d997fc0429fe2199ebf6637a52943 Mon Sep 17 00:00:00 2001 From: Orion Date: Fri, 9 Jan 2026 00:23:22 +0000 Subject: [PATCH] :bookmark: Bump ik-llama.cpp-cuda version to r4099.145e4f4e-1 --- .SRCINFO | 2 +- PKGBUILD | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/.SRCINFO b/.SRCINFO index 2316bf9eb7665..74a375e5f2b88 100644 --- a/.SRCINFO +++ b/.SRCINFO @@ -1,6 +1,6 @@ pkgbase = ik-llama.cpp-cuda pkgdesc = llama.cpp fork with additional SOTA quants and improved performance (CUDA Backend) - pkgver = r4096.1b241928 + pkgver = r4099.145e4f4e pkgrel = 1 url = https://github.com/ikawrakow/ik_llama.cpp arch = x86_64 diff --git a/PKGBUILD b/PKGBUILD index 1858687b45161..37e0e49603f01 100644 --- a/PKGBUILD +++ b/PKGBUILD @@ -2,7 +2,7 @@ pkgname=ik-llama.cpp-cuda _pkgname=ik_llama.cpp -pkgver=r4096.1b241928 +pkgver=r4099.145e4f4e pkgrel=1 pkgdesc="llama.cpp fork with additional SOTA quants and improved performance (CUDA Backend)" arch=(x86_64 armv7h aarch64)