llm: auto detect models that require Ollama Engine (#1)

This commit is contained in:
Daniel Hiltgen 2025-03-11 04:25:16 -07:00 committed by Michael Yang
parent 11bfa62796
commit ab39e08eb9
2 changed files with 5 additions and 1 deletions

View file

@ -271,7 +271,7 @@ func NewLlamaServer(gpus discover.GpuInfoList, modelPath string, f *ggml.GGML, a
var llamaModel *llama.Model
var textProcessor model.TextProcessor
if envconfig.NewEngine() {
if envconfig.NewEngine() || f.KV().OllamaEngineRequired() {
textProcessor, err = model.NewTextProcessor(modelPath)
if err != nil {
// To prepare for opt-out mode, instead of treating this as an error, we fallback to the old runner