mirror of
https://github.com/ollama/ollama.git
synced 2025-05-16 22:44:25 +02:00
We currently preallocate compute graph memory for the worst case batch of text tokens. This adds support for doing the same for images. Note that image models are more complicated than text models in how they process their inputs so there may be cases where this approach isn't completely generic for all models. It covers all currently supported models though. |
||
---|---|---|
.. | ||
cache.go | ||
cache_test.go | ||
multimodal.go | ||
runner.go |