mirror of
https://github.com/ollama/ollama.git
synced 2025-05-10 18:06:33 +02:00
* Move quantization logic to GGML via new backend This moves the model aware logic to Go code and calls GGMLs quantization code for model creation. * Remove "add model quantizations" This is no longer needed now that quantization is implemented in Go+GGML code directly. |
||
---|---|---|
.. | ||
sentencepiece | ||
testdata | ||
convert.go | ||
convert_bert.go | ||
convert_commandr.go | ||
convert_gemma.go | ||
convert_gemma2.go | ||
convert_gemma2_adapter.go | ||
convert_gemma3.go | ||
convert_llama.go | ||
convert_llama4.go | ||
convert_llama_adapter.go | ||
convert_mistral.go | ||
convert_mixtral.go | ||
convert_phi3.go | ||
convert_qwen2.go | ||
convert_test.go | ||
fs.go | ||
reader.go | ||
reader_safetensors.go | ||
reader_torch.go | ||
sentencepiece_model.proto | ||
tokenizer.go | ||
tokenizer_spm.go | ||
tokenizer_test.go |