# Generated by abuild 3.17.0_rc2-r0
# using fakeroot version 1.37.2
pkgname = llama.cpp-libs
pkgver = 0.0.9006-r0
pkgdesc = LLM inference in C/C++ (with Vulkan GPU acceleration) (shared libraries)
url = https://github.com/ggml-org/llama.cpp
builddate = 1777872529
packager = Buildozer <alpine-devel@lists.alpinelinux.org>
size = 10341800
arch = s390x
origin = llama.cpp
commit = a36270a0145ca7c911b50944793667f2d3e41319
maintainer = Hugo Osvaldo Barrera <hugo@whynothugo.nl>
license = MIT
# automatically detected:
provides = so:llama.cpp:libggml-base.so.0=0.10.2
provides = so:llama.cpp:libggml-blas.so=0
provides = so:llama.cpp:libggml-rpc.so=0
provides = so:llama.cpp:libggml.so.0=0.10.2
provides = so:llama.cpp:libllama-common.so.0=0.0.9006
provides = so:llama.cpp:libllama.so.0=0.0.9006
provides = so:llama.cpp:libmtmd.so.0=0.0.9006
depend = openblas
depend = so:libc.musl-s390x.so.1
depend = so:libcrypto.so.3
depend = so:libgcc_s.so.1
depend = so:libssl.so.3
depend = so:libstdc++.so.6
datahash = ac462b73e5b94dae1e7a6c3ecb3b9fe5c4cc28377bb30cc00dc41e71bb78855f
