# Generated by makepkg 6.1.0
# Sat Mar 15 09:10:57 UTC 2025
pkgbase = mingw-w64-llama.cpp
	pkgdesc = Library and tools for running inference with Meta's LLaMA model (and derivatives) in C/C++ (mingw-w64)
	pkgver = b4889
	pkgrel = 1
	epoch = 1
	url = https://github.com/ggml-org/llama.cpp
	arch = any
	license = spdx:MIT
	makedepends = mingw-w64-ucrt-x86_64-cc
	makedepends = mingw-w64-ucrt-x86_64-cmake
	makedepends = mingw-w64-ucrt-x86_64-ninja
	makedepends = mingw-w64-ucrt-x86_64-opencl-headers
	makedepends = mingw-w64-ucrt-x86_64-shaderc
	makedepends = mingw-w64-ucrt-x86_64-vulkan-headers
	depends = mingw-w64-ucrt-x86_64-curl
	depends = mingw-w64-ucrt-x86_64-gcc-libs
	depends = mingw-w64-ucrt-x86_64-omp
	depends = mingw-w64-ucrt-x86_64-opencl-icd
	depends = mingw-w64-ucrt-x86_64-openblas
	depends = mingw-w64-ucrt-x86_64-python
	depends = mingw-w64-ucrt-x86_64-python-numpy
	depends = mingw-w64-ucrt-x86_64-vulkan-loader
	source = https://github.com/ggml-org/llama.cpp/archive/b4889/llama.cpp-b4889.tar.gz
	sha256sums = ab1cf3d8f913534bf8d35bf1bd93db45d1692654ee8c14799d415661d15f9848

pkgname = mingw-w64-ucrt-x86_64-llama.cpp
