ollama/llm
Daniel Hiltgen 424810450f
Move quantization to new backend (#10363)
* Move quantization logic to GGML via new backend

This moves the model aware logic to Go code and calls GGMLs quantization code for model creation.

* Remove "add model quantizations"

This is no longer needed now that quantization is implemented in Go+GGML code directly.
2025-05-06 11:20:48 -07:00
..
llm_darwin.go Optimize container images for startup (#6547) 2024-09-12 12:10:30 -07:00
llm_linux.go Optimize container images for startup (#6547) 2024-09-12 12:10:30 -07:00
llm_windows.go win: lint fix (#10571) 2025-05-05 11:08:12 -07:00
memory.go explicitly decode maxarraysize 1024 2025-04-25 16:59:01 -07:00
memory_test.go Move quantization to new backend (#10363) 2025-05-06 11:20:48 -07:00
server.go api: remove unused or unsupported api options (#10574) 2025-05-05 14:54:40 -07:00
server_test.go llm: do not error on "null" format (#8139) 2024-12-17 09:49:37 -08:00
status.go Improve crash reporting (#7728) 2024-11-19 16:26:57 -08:00