mirror of
https://github.com/archlinux/aur.git
synced 2026-03-14 23:16:48 +01:00
upgpkg: llama.cpp-vulkan b5669-2
Signed-off-by: txtsd <code@ihavea.quest>
This commit is contained in:
parent
db5c71d849
commit
f09097f74c
2 changed files with 8 additions and 36 deletions
13
.SRCINFO
13
.SRCINFO
|
|
@ -1,7 +1,7 @@
|
|||
pkgbase = llama.cpp-vulkan
|
||||
pkgdesc = Port of Facebook's LLaMA model in C/C++ (with Vulkan GPU optimizations)
|
||||
pkgver = b5669
|
||||
pkgrel = 1
|
||||
pkgrel = 2
|
||||
url = https://github.com/ggerganov/llama.cpp
|
||||
arch = x86_64
|
||||
arch = armv7h
|
||||
|
|
@ -9,29 +9,22 @@ pkgbase = llama.cpp-vulkan
|
|||
license = MIT
|
||||
makedepends = cmake
|
||||
makedepends = git
|
||||
makedepends = shaderc
|
||||
makedepends = vulkan-headers
|
||||
makedepends = pkgconf
|
||||
depends = blas-openblas
|
||||
depends = blas64-openblas
|
||||
depends = curl
|
||||
depends = gcc-libs
|
||||
depends = glibc
|
||||
depends = openmp
|
||||
depends = libggml-vulkan
|
||||
depends = python
|
||||
depends = python-numpy
|
||||
depends = python-sentencepiece
|
||||
depends = vulkan-icd-loader
|
||||
optdepends = python-pytorch
|
||||
provides = llama.cpp
|
||||
conflicts = llama.cpp
|
||||
options = lto
|
||||
options = !debug
|
||||
source = git+https://github.com/ggerganov/llama.cpp#tag=b5669
|
||||
source = git+https://github.com/nomic-ai/kompute.git
|
||||
source = llama.cpp.conf
|
||||
source = llama.cpp.service
|
||||
sha256sums = eb46ae341d2f89395c4c86186c25be504a9e4db3369e1a826ccb3bc50743d21c
|
||||
sha256sums = SKIP
|
||||
sha256sums = 53fa70cfe40cb8a3ca432590e4f76561df0f129a31b121c9b4b34af0da7c4d87
|
||||
sha256sums = 0377d08a07bda056785981d3352ccd2dbc0387c4836f91fb73e6b790d836620d
|
||||
|
||||
|
|
|
|||
31
PKGBUILD
31
PKGBUILD
|
|
@ -3,67 +3,47 @@
|
|||
pkgname=llama.cpp-vulkan
|
||||
_pkgname=${pkgname%%-vulkan}
|
||||
pkgver=b5669
|
||||
pkgrel=1
|
||||
pkgrel=2
|
||||
pkgdesc="Port of Facebook's LLaMA model in C/C++ (with Vulkan GPU optimizations)"
|
||||
arch=(x86_64 armv7h aarch64)
|
||||
url='https://github.com/ggerganov/llama.cpp'
|
||||
license=('MIT')
|
||||
depends=(
|
||||
blas-openblas
|
||||
blas64-openblas
|
||||
curl
|
||||
gcc-libs
|
||||
glibc
|
||||
openmp
|
||||
libggml-vulkan
|
||||
python
|
||||
python-numpy
|
||||
python-sentencepiece
|
||||
vulkan-icd-loader
|
||||
)
|
||||
makedepends=(
|
||||
cmake
|
||||
git
|
||||
shaderc
|
||||
vulkan-headers
|
||||
pkgconf
|
||||
)
|
||||
optdepends=(python-pytorch)
|
||||
provides=(${_pkgname})
|
||||
conflicts=(${_pkgname})
|
||||
options+=(lto)
|
||||
options=(lto !debug)
|
||||
source=(
|
||||
"git+${url}#tag=${pkgver}"
|
||||
"git+https://github.com/nomic-ai/kompute.git"
|
||||
llama.cpp.conf
|
||||
llama.cpp.service
|
||||
)
|
||||
sha256sums=('eb46ae341d2f89395c4c86186c25be504a9e4db3369e1a826ccb3bc50743d21c'
|
||||
'SKIP'
|
||||
'53fa70cfe40cb8a3ca432590e4f76561df0f129a31b121c9b4b34af0da7c4d87'
|
||||
'0377d08a07bda056785981d3352ccd2dbc0387c4836f91fb73e6b790d836620d')
|
||||
|
||||
prepare() {
|
||||
cd "${_pkgname}"
|
||||
|
||||
git submodule init
|
||||
git config submodule.kompute.url "${srcdir}/kompute"
|
||||
git -c protocol.file.allow=always submodule update
|
||||
}
|
||||
|
||||
build() {
|
||||
local _cmake_options=(
|
||||
-B build
|
||||
-S "${_pkgname}"
|
||||
-DCMAKE_BUILD_TYPE=None
|
||||
-DCMAKE_INSTALL_PREFIX='/usr'
|
||||
-DGGML_ALL_WARNINGS=OFF
|
||||
-DGGML_ALL_WARNINGS_3RD_PARTY=OFF
|
||||
-DBUILD_SHARED_LIBS=ON
|
||||
-DGGML_LTO=ON
|
||||
-DGGML_RPC=ON
|
||||
-DLLAMA_CURL=ON
|
||||
-DGGML_BLAS=ON
|
||||
-DGGML_VULKAN=ON
|
||||
-DLLAMA_BUILD_TESTS=OFF
|
||||
-DLLAMA_USE_SYSTEM_GGML=ON
|
||||
-Wno-dev
|
||||
)
|
||||
cmake "${_cmake_options[@]}"
|
||||
|
|
@ -72,7 +52,6 @@ build() {
|
|||
|
||||
package() {
|
||||
DESTDIR="${pkgdir}" cmake --install build
|
||||
rm "${pkgdir}/usr/include/"ggml*
|
||||
|
||||
install -Dm644 "${_pkgname}/LICENSE" "${pkgdir}/usr/share/licenses/${pkgname}/LICENSE"
|
||||
|
||||
|
|
|
|||
Loading…
Add table
Reference in a new issue