mirror of
https://github.com/ollama/ollama
synced 2026-04-23 08:45:14 +00:00
Match the ollamarunner and OpenAI semantics: raw, full-vocab log-softmax with the top-K ranked by probability. Skipped on the GPU when the request doesn't ask for logprobs so decode doesn't pay for it otherwise. |
||
|---|---|---|
| .. | ||
| cache | ||
| mlx | ||
| model | ||
| sample | ||
| cache.go | ||
| cache_test.go | ||
| cache_trie.go | ||
| cache_trie_test.go | ||
| client.go | ||
| imports.go | ||
| pipeline.go | ||
| runner.go | ||
| server.go | ||
| utf8_buffer.go | ||
| utf8_buffer_test.go | ||