From adb759b3524c7159ce00b06cb5bf7ad13a059910 Mon Sep 17 00:00:00 2001 From: Orion Date: Thu, 5 Feb 2026 16:39:03 +0000 Subject: [PATCH] :bookmark: Bump ik-llama.cpp-cuda version to r4180.1ec12b8e-1 --- .SRCINFO | 2 +- PKGBUILD | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/.SRCINFO b/.SRCINFO index 05e2bd94bcaeb..6a34ab457ec44 100644 --- a/.SRCINFO +++ b/.SRCINFO @@ -1,6 +1,6 @@ pkgbase = ik-llama.cpp-cuda pkgdesc = llama.cpp fork with additional SOTA quants and improved performance (CUDA Backend) - pkgver = r4179.51fc7875 + pkgver = r4180.1ec12b8e pkgrel = 1 url = https://github.com/ikawrakow/ik_llama.cpp arch = x86_64 diff --git a/PKGBUILD b/PKGBUILD index c21921b060607..85a01bdbfbff2 100644 --- a/PKGBUILD +++ b/PKGBUILD @@ -2,7 +2,7 @@ pkgname=ik-llama.cpp-cuda _pkgname=ik_llama.cpp -pkgver=r4179.51fc7875 +pkgver=r4180.1ec12b8e pkgrel=1 pkgdesc="llama.cpp fork with additional SOTA quants and improved performance (CUDA Backend)" arch=(x86_64 armv7h aarch64)