ollama/model/models
Daniel Hiltgen 87288ced4f
New models (#15861)
* mlx: add laguna model support

* convert: support fp8 safetensors import

Decode HF F8_E4M3 safetensors with block scale companions into GGUF-supported tensor types, and record which output tensors came from FP8 source weights.

Use that source-precision metadata during create quantization: default FP8-sourced GGUFs to Q8_0, keep non-FP8 tensors at their original precision for Q8_0, and promote non-FP8 quantizable tensors to Q8_0 for Q4_K requests.

* ggml: add laguna model support

* server: preserve generate logprobs with builtin parsers

Generate requests were dropping logprob-only chunks whenever a builtin parser buffered visible content. Chat already handled this case, but generate only forwarded chunks with visible response, thinking, or tool-call output.

Keep generate chunks that carry logprobs even when the builtin parser has not flushed visible content yet, and add a regression test that exercises the behavior with a generic thinking parser.

* review comments - perf improvements

* ggml: implement nemotron 3 nano omni

* add poolside integration

* update poolside doc

* adapt to new cache setup

* fix test

* fix test

---------

Co-authored-by: Eva Ho <hoyyeva@gmail.com>
2026-04-28 11:50:12 -07:00
..
bert move tokenizers to separate package (#13825) 2026-02-05 17:44:11 -08:00
deepseek2 move tokenizers to separate package (#13825) 2026-02-05 17:44:11 -08:00
deepseekocr move tokenizers to separate package (#13825) 2026-02-05 17:44:11 -08:00
gemma2 move tokenizers to separate package (#13825) 2026-02-05 17:44:11 -08:00
gemma3 move tokenizers to separate package (#13825) 2026-02-05 17:44:11 -08:00
gemma3n move tokenizers to separate package (#13825) 2026-02-05 17:44:11 -08:00
gemma4 tokenizer: add byte fallback for SentencePiece BPE encoding (#15232) 2026-04-02 13:04:45 -07:00
glm4moelite move tokenizers to separate package (#13825) 2026-02-05 17:44:11 -08:00
glmocr move tokenizers to separate package (#13825) 2026-02-05 17:44:11 -08:00
gptoss move tokenizers to separate package (#13825) 2026-02-05 17:44:11 -08:00
laguna New models (#15861) 2026-04-28 11:50:12 -07:00
lfm2 model: improvements to LFM architectures (#14368) 2026-02-23 14:38:10 -08:00
llama move tokenizers to separate package (#13825) 2026-02-05 17:44:11 -08:00
llama4 move tokenizers to separate package (#13825) 2026-02-05 17:44:11 -08:00
mistral3 move tokenizers to separate package (#13825) 2026-02-05 17:44:11 -08:00
mllama move tokenizers to separate package (#13825) 2026-02-05 17:44:11 -08:00
nemotronh New models (#15861) 2026-04-28 11:50:12 -07:00
nomicbert move tokenizers to separate package (#13825) 2026-02-05 17:44:11 -08:00
olmo3 move tokenizers to separate package (#13825) 2026-02-05 17:44:11 -08:00
qwen2 move tokenizers to separate package (#13825) 2026-02-05 17:44:11 -08:00
qwen3 move tokenizers to separate package (#13825) 2026-02-05 17:44:11 -08:00
qwen3next model: add qwen3-next compatibility for legacy ssm_in projections (#15133) 2026-03-29 11:50:47 -07:00
qwen3vl model: support for qwen3.5 architecture (#14378) 2026-02-24 20:08:05 -08:00
qwen25vl move tokenizers to separate package (#13825) 2026-02-05 17:44:11 -08:00
models.go New models (#15861) 2026-04-28 11:50:12 -07:00