ollama/template
Michael Yang 58245413f4
next ollama runner (#7913)
feat: add new Ollama engine using ggml through cgo

This change introduces a new way to run pretrained models. It introduces 3 high level interfaces and a bunch of smaller helper interfaces to facilitate this.

- `model.Model` defines the interface for a model architecture. Models such as `llama` and `mllama`, which are provided as examples, can implement the model's forward propagation in the `Forward` method. This method will be called to generate completions. This interface can be found in `model/model.go`
- `ml.Backend` defines the interface for a backend tensor library, in this case `ggml`. Among other things, a Backend is responsible for loading a pretrained model into hardware (GPU, CPU, etc) and providing an interface for Models to access loaded tensors. This interface can be found in `ml/backend.go`
- `ml.Tensor` defines the interface for a tensor and tensor operations

This is the first implementation of the new engine. Follow up PRs will implement more features:

- non-greedy sampling (#8410)
- integration with Ollama and KV caching (#8301)
- more model support (#9080) with more coming soon

Co-authored-by: Bruce MacDonald <brucewmacdonald@gmail.com>
2025-02-13 16:31:21 -08:00
..
testdata convert: import support for command-r models from safetensors (#6063) 2025-01-15 16:31:22 -08:00
alfred.gotmpl update templates to use messages 2024-08-27 15:44:04 -07:00
alfred.json autodetect stop parameters from template 2024-07-12 16:01:23 -07:00
alpaca.gotmpl update templates to use messages 2024-08-27 15:44:04 -07:00
alpaca.json autodetect stop parameters from template 2024-07-12 16:01:23 -07:00
chatml.gotmpl update templates to use messages 2024-08-27 15:44:04 -07:00
chatml.json autodetect stop parameters from template 2024-07-12 16:01:23 -07:00
chatqa.gotmpl update templates to use messages 2024-08-27 15:44:04 -07:00
chatqa.json autodetect stop parameters from template 2024-07-12 16:01:23 -07:00
codellama-70b-instruct.gotmpl update templates to use messages 2024-08-27 15:44:04 -07:00
codellama-70b-instruct.json autodetect stop parameters from template 2024-07-12 16:01:23 -07:00
command-r.gotmpl convert: import support for command-r models from safetensors (#6063) 2025-01-15 16:31:22 -08:00
command-r.json convert: import support for command-r models from safetensors (#6063) 2025-01-15 16:31:22 -08:00
falcon-instruct.gotmpl update templates to use messages 2024-08-27 15:44:04 -07:00
falcon-instruct.json autodetect stop parameters from template 2024-07-12 16:01:23 -07:00
gemma-instruct.gotmpl update templates to use messages 2024-08-27 15:44:04 -07:00
gemma-instruct.json autodetect stop parameters from template 2024-07-12 16:01:23 -07:00
granite-instruct.gotmpl update templates to use messages 2024-08-27 15:44:04 -07:00
granite-instruct.json autodetect stop parameters from template 2024-07-12 16:01:23 -07:00
index.json convert: import support for command-r models from safetensors (#6063) 2025-01-15 16:31:22 -08:00
llama2-chat.gotmpl update templates to use messages 2024-08-27 15:44:04 -07:00
llama2-chat.json autodetect stop parameters from template 2024-07-12 16:01:23 -07:00
llama3-instruct.gotmpl update templates to use messages 2024-08-27 15:44:04 -07:00
llama3-instruct.json autodetect stop parameters from template 2024-07-12 16:01:23 -07:00
magicoder.gotmpl update templates to use messages 2024-08-27 15:44:04 -07:00
magicoder.json autodetect stop parameters from template 2024-07-12 16:01:23 -07:00
mistral-instruct.gotmpl update templates to use messages 2024-08-27 15:44:04 -07:00
mistral-instruct.json autodetect stop parameters from template 2024-07-12 16:01:23 -07:00
openchat.gotmpl update templates to use messages 2024-08-27 15:44:04 -07:00
openchat.json autodetect stop parameters from template 2024-07-12 16:01:23 -07:00
phi-3.gotmpl update templates to use messages 2024-08-27 15:44:04 -07:00
phi-3.json autodetect stop parameters from template 2024-07-12 16:01:23 -07:00
solar-instruct.gotmpl update templates to use messages 2024-08-27 15:44:04 -07:00
solar-instruct.json autodetect stop parameters from template 2024-07-12 16:01:23 -07:00
starcoder2-instruct.gotmpl update templates to use messages 2024-08-27 15:44:04 -07:00
starcoder2-instruct.json autodetect stop parameters from template 2024-07-12 16:01:23 -07:00
template.go image processing for llama3.2 (#6963) 2024-10-18 16:12:35 -07:00
template_test.go next ollama runner (#7913) 2025-02-13 16:31:21 -08:00
vicuna.gotmpl update templates to use messages 2024-08-27 15:44:04 -07:00
vicuna.json autodetect stop parameters from template 2024-07-12 16:01:23 -07:00
zephyr.gotmpl update templates to use messages 2024-08-27 15:44:04 -07:00
zephyr.json autodetect stop parameters from template 2024-07-12 16:01:23 -07:00