%FILENAME%
llama.cpp-b7677-1-x86_64.pkg.tar.zst

%NAME%
llama.cpp

%BASE%
llama.cpp

%VERSION%
b7677-1

%DESC%
Port of Facebook's LLaMA model in C/C++ (with system ggml support)

%CSIZE%
6761211

%ISIZE%
28229141

%SHA256SUM%
c1993bd811f0da4650039fe3e6c19bc9b365baf9a189b17c939453af22961f04

%URL%
https://github.com/ggerganov/llama.cpp

%LICENSE%
MIT

%ARCH%
x86_64

%BUILDDATE%
1767905523

%PACKAGER%
lilac (on behalf of 依云) <lilydjwg@gmail.com>

%DEPENDS%
curl
gcc-libs
python
ggml

%OPTDEPENDS%
python-numpy: needed for convert_hf_to_gguf.py
python-safetensors: needed for convert_hf_to_gguf.py
python-sentencepiece: needed for convert_hf_to_gguf.py
python-pytorch: needed for convert_hf_to_gguf.py
python-transformers: needed for convert_hf_to_gguf.py

%MAKEDEPENDS%
cmake
git
vulkan-radeon

