From 2b564ef76a6ab8b758dac4ac59ad8e52fff26f62 Mon Sep 17 00:00:00 2001 From: Orion Date: Wed, 17 Dec 2025 05:18:10 +0000 Subject: [PATCH] :bookmark: Bump llama.cpp-vulkan version to b7441-1 --- .SRCINFO | 7 ++++--- PKGBUILD | 5 +++-- 2 files changed, 7 insertions(+), 5 deletions(-) diff --git a/.SRCINFO b/.SRCINFO index 4ad80b14e5f0f..55debfb40c5d7 100644 --- a/.SRCINFO +++ b/.SRCINFO @@ -1,6 +1,6 @@ pkgbase = llama.cpp-vulkan pkgdesc = Port of Facebook's LLaMA model in C/C++ (with Vulkan GPU optimizations) - pkgver = b7437 + pkgver = b7441 pkgrel = 1 url = https://github.com/ggml-org/llama.cpp arch = x86_64 @@ -28,10 +28,11 @@ pkgbase = llama.cpp-vulkan conflicts = stable-diffusion.cpp options = lto options = !debug - source = llama.cpp-vulkan-b7437.tar.gz::https://github.com/ggml-org/llama.cpp/archive/refs/tags/b7437.tar.gz + backup = etc/conf.d/llama.cpp + source = llama.cpp-vulkan-b7441.tar.gz::https://github.com/ggml-org/llama.cpp/archive/refs/tags/b7441.tar.gz source = https://raw.githubusercontent.com/Orion-zhen/aur-packages/refs/heads/main/assets/llama.cpp/llama.cpp.service source = https://raw.githubusercontent.com/Orion-zhen/aur-packages/refs/heads/main/assets/llama.cpp/llama.cpp.conf - sha256sums = 13fc729656c94b3b44e9c0cf9531087094e550a826243a953e76d62cfecefac2 + sha256sums = 60f8f117308584b5004e4913cfc096e02e3be6d21d9f3e1cac15781c10f5ffeb sha256sums = 0377d08a07bda056785981d3352ccd2dbc0387c4836f91fb73e6b790d836620d sha256sums = e4856f186f69cd5dbfcc4edec9f6b6bd08e923bceedd8622eeae1a2595beb2ec diff --git a/PKGBUILD b/PKGBUILD index e3a97c886aaa4..04e12cd4ae4c2 100644 --- a/PKGBUILD +++ b/PKGBUILD @@ -3,7 +3,7 @@ pkgname=llama.cpp-vulkan _pkgname=${pkgname%%-vulkan} -pkgver=b7437 +pkgver=b7441 pkgrel=1 pkgdesc="Port of Facebook's LLaMA model in C/C++ (with Vulkan GPU optimizations)" arch=(x86_64 armv7h aarch64) @@ -32,12 +32,13 @@ optdepends=( provides=(${_pkgname}) conflicts=(${_pkgname} libggml ggml stable-diffusion.cpp) options=(lto !debug) +backup=("etc/conf.d/llama.cpp") source=( "${pkgname}-${pkgver}.tar.gz::https://github.com/ggml-org/llama.cpp/archive/refs/tags/${pkgver}.tar.gz" "https://raw.githubusercontent.com/Orion-zhen/aur-packages/refs/heads/main/assets/llama.cpp/llama.cpp.service" "https://raw.githubusercontent.com/Orion-zhen/aur-packages/refs/heads/main/assets/llama.cpp/llama.cpp.conf" ) -sha256sums=('13fc729656c94b3b44e9c0cf9531087094e550a826243a953e76d62cfecefac2' +sha256sums=('60f8f117308584b5004e4913cfc096e02e3be6d21d9f3e1cac15781c10f5ffeb' '0377d08a07bda056785981d3352ccd2dbc0387c4836f91fb73e6b790d836620d' 'e4856f186f69cd5dbfcc4edec9f6b6bd08e923bceedd8622eeae1a2595beb2ec')