ollama/template
Daniel Hiltgen 31e336791a runner: Remove CGO engines, use llama-server exclusively for GGML models
Remove the vendored GGML and llama.cpp backend, CGO runner, Go model
implementations, and sample.  llama-server (built from upstream llama.cpp via
FetchContent) is now the sole inference engine for GGUF-based models.
(Safetensor based models continue to run on the new MLX engine.)  This allows
us to more rapidly pick up new capabilities and fixes from llama.cpp as they
come out.

On windows this now requires recent AMD driver versions to support ROCm v7 as
llama.cpp currently does not support building against v6.
2026-05-06 17:26:05 -07:00
..
testdata templates: add autotemplate for gemma3 (#9880) 2025-03-20 00:15:30 -07:00
alfred.gotmpl
alfred.json
alpaca.gotmpl
alpaca.json
chatml.gotmpl
chatml.json
chatqa.gotmpl
chatqa.json
codellama-70b-instruct.gotmpl
codellama-70b-instruct.json
command-r.gotmpl
command-r.json
falcon-instruct.gotmpl
falcon-instruct.json
gemma-instruct.gotmpl
gemma-instruct.json
gemma3-instruct.gotmpl templates: add autotemplate for gemma3 (#9880) 2025-03-20 00:15:30 -07:00
gemma3-instruct.json templates: add autotemplate for gemma3 (#9880) 2025-03-20 00:15:30 -07:00
granite-instruct.gotmpl
granite-instruct.json
index.json templates: add autotemplate for gemma3 (#9880) 2025-03-20 00:15:30 -07:00
llama2-chat.gotmpl
llama2-chat.json
llama3-instruct.gotmpl
llama3-instruct.json
magicoder.gotmpl
magicoder.json
mistral-instruct.gotmpl
mistral-instruct.json
openchat.gotmpl
openchat.json
phi-3.gotmpl
phi-3.json
solar-instruct.gotmpl
solar-instruct.json
starcoder2-instruct.gotmpl
starcoder2-instruct.json
template.go runner: Remove CGO engines, use llama-server exclusively for GGML models 2026-05-06 17:26:05 -07:00
template_test.go template: fix args-as-json rendering (#13636) 2026-01-06 18:33:57 -08:00
vicuna.gotmpl
vicuna.json
zephyr.gotmpl
zephyr.json