%FILENAME%
llama.cpp-b7548-1-x86_64.pkg.tar.zst

%NAME%
llama.cpp

%BASE%
llama.cpp

%VERSION%
b7548-1

%DESC%
Port of Facebook's LLaMA model in C/C++ (with system ggml support)

%CSIZE%
6960408

%ISIZE%
29323318

%SHA256SUM%
fb1c975597e43ba2bfb7751f659b86f3fefd58c22cbfdaf252fbd042b87bf612

%URL%
https://github.com/ggerganov/llama.cpp

%LICENSE%
MIT

%ARCH%
x86_64

%BUILDDATE%
1766782084

%PACKAGER%
lilac (on behalf of 依云) <lilydjwg@gmail.com>

%DEPENDS%
curl
gcc-libs
python
ggml

%OPTDEPENDS%
python-numpy: needed for convert_hf_to_gguf.py
python-safetensors: needed for convert_hf_to_gguf.py
python-sentencepiece: needed for convert_hf_to_gguf.py
python-pytorch: needed for convert_hf_to_gguf.py
python-transformers: needed for convert_hf_to_gguf.py

%MAKEDEPENDS%
cmake
git
vulkan-radeon

