ollama/ml
2026-01-23 20:10:04 -08:00
..
backend llama: fix CUDA MMA errors in release build (#13874) 2026-01-23 20:10:04 -08:00
nn fix: qwen2.5 vl rope (#13486) 2025-12-15 17:30:33 -08:00
backend.go model: add lfm2 architecture and LFM2.5-1.2B-Thinking support (#13792) 2026-01-20 12:20:53 -08:00
device.go flash attn: add auto mode for llama engine (#13052) 2025-12-12 13:27:19 -08:00
path.go