ollama/fs
Jesse Gross 19e6796eac llm: Support KV cache quantization with gpt-oss
With the new version of GGML in #12245, KV cache quantization
no longer causes a fallback to CPU.
2025-10-03 16:31:58 -07:00
..
ggml llm: Support KV cache quantization with gpt-oss 2025-10-03 16:31:58 -07:00
gguf Reapply "feat: incremental gguf parser (#10822)" (#11114) (#11119) 2025-06-20 11:11:40 -07:00
util/bufioutil next ollama runner (#7913) 2025-02-13 16:31:21 -08:00
config.go add new gemma model (#11204) 2025-06-25 21:47:09 -07:00