From 1ba2c7e6bf5d02c68b853a84c461f028226b11b5 Mon Sep 17 00:00:00 2001 From: Orion Date: Sat, 17 Jan 2026 07:04:46 +0000 Subject: [PATCH] :bookmark: Bump llama.cpp-vulkan version to b7761-1 --- .SRCINFO | 6 +++--- PKGBUILD | 6 +++--- 2 files changed, 6 insertions(+), 6 deletions(-) diff --git a/.SRCINFO b/.SRCINFO index 4842218285ba6..6489e71123017 100644 --- a/.SRCINFO +++ b/.SRCINFO @@ -1,6 +1,6 @@ pkgbase = llama.cpp-vulkan pkgdesc = Port of Facebook's LLaMA model in C/C++ (with Vulkan GPU optimizations) - pkgver = b7760 + pkgver = b7761 pkgrel = 1 url = https://github.com/ggml-org/llama.cpp arch = x86_64 @@ -30,10 +30,10 @@ pkgbase = llama.cpp-vulkan options = lto options = !debug backup = etc/conf.d/llama.cpp - source = llama.cpp-vulkan-b7760.tar.gz::https://github.com/ggml-org/llama.cpp/archive/refs/tags/b7760.tar.gz + source = llama.cpp-vulkan-b7761.tar.gz::https://github.com/ggml-org/llama.cpp/archive/refs/tags/b7761.tar.gz source = https://raw.githubusercontent.com/Orion-zhen/aur-packages/refs/heads/main/assets/llama.cpp/llama.cpp.service source = https://raw.githubusercontent.com/Orion-zhen/aur-packages/refs/heads/main/assets/llama.cpp/llama.cpp.conf - sha256sums = 4053d9aec22cfad8c06610d4092adafd47bff340b930f40c99028e9e11c39675 + sha256sums = 33c4185b529a377377ea6989cf2be6871c6517d229d06227a3120d918a98e600 sha256sums = 0377d08a07bda056785981d3352ccd2dbc0387c4836f91fb73e6b790d836620d sha256sums = e4856f186f69cd5dbfcc4edec9f6b6bd08e923bceedd8622eeae1a2595beb2ec diff --git a/PKGBUILD b/PKGBUILD index b759bc8e17a86..1e78f4cca4092 100644 --- a/PKGBUILD +++ b/PKGBUILD @@ -3,7 +3,7 @@ pkgname=llama.cpp-vulkan _pkgname=${pkgname%%-vulkan} -pkgver=b7760 +pkgver=b7761 pkgrel=1 pkgdesc="Port of Facebook's LLaMA model in C/C++ (with Vulkan GPU optimizations)" arch=(x86_64 armv7h aarch64) @@ -39,7 +39,7 @@ source=( "https://raw.githubusercontent.com/Orion-zhen/aur-packages/refs/heads/main/assets/llama.cpp/llama.cpp.service" "https://raw.githubusercontent.com/Orion-zhen/aur-packages/refs/heads/main/assets/llama.cpp/llama.cpp.conf" ) -sha256sums=('4053d9aec22cfad8c06610d4092adafd47bff340b930f40c99028e9e11c39675' +sha256sums=('33c4185b529a377377ea6989cf2be6871c6517d229d06227a3120d918a98e600' '0377d08a07bda056785981d3352ccd2dbc0387c4836f91fb73e6b790d836620d' 'e4856f186f69cd5dbfcc4edec9f6b6bd08e923bceedd8622eeae1a2595beb2ec') @@ -76,7 +76,7 @@ build() { -DGGML_BACKEND_DL=ON -DGGML_CPU_ALL_VARIANTS=ON -DGGML_NATIVE=OFF - -DGGML_ZENDNN=ON + # -DGGML_ZENDNN=ON ) else # 本地构建, 针对当前设备优化