ollama/model
Devon Rifkin bf2a421727
gemma4: restore e2b-style nothink prompt (#15560)
Gemma 4 prompts differ when thinking is disabled for different sized
models: 26b/31b emit an empty thought block, while e2b/e4b do not.

Before #15490, our shared Gemma 4 renderer effectively matched the
e2b behavior. #15490 changed it to always emit the empty thought block,
which regressed e2b/e4b nothink behavior and led to #15536 (and possibly

This change restores the previous shared behavior by removing the empty
trailing thought block. It also renames the checked-in upstream chat
templates so the e2b and 31b fixtures are tracked separately.

A follow-up will split Gemma 4 rendering by model size.

Fixes: #15536
2026-04-13 14:26:15 -07:00
..
imageproc deepseekocr 2025-11-18 16:11:37 -08:00
input batch: use tensors for outputs (#12185) 2025-09-15 14:33:06 -07:00
models tokenizer: add byte fallback for SentencePiece BPE encoding (#15232) 2026-04-02 13:04:45 -07:00
parsers gemma4: be less strict about whitespace before bare keys (#15494) 2026-04-11 16:30:27 -07:00
renderers gemma4: restore e2b-style nothink prompt (#15560) 2026-04-13 14:26:15 -07:00
model.go Add support for gemma4 (#15214) 2026-04-02 11:33:33 -07:00
model_test.go Add support for gemma4 (#15214) 2026-04-02 11:33:33 -07:00