mirror of
https://github.com/ollama/ollama.git
synced 2025-05-11 02:16:36 +02:00
- Allows specifying whether thinking mode should be on or not - Templates get passed a new option so, e.g., qwen3's template can put `/think` or `/no_think` in the system prompt depending on the value of the setting - Add parsing for thinking blocks in both streaming/non-streaming mode - Update the CLI to make use of these changes TODO: - [ ] Don't parse thinking blocks when the user doesn't explicitly set the option, to maintain backwards compatibility - [ ] Warning on CLI when using a non-thinking/older version of a model (with an old template) - [ ] Wire up capabilities fully - [x] Unify parsing for streaming/non-streaming - [ ] Update templates - [ ] Update python/js libraries - [ ] How to handle differences in models wrt defaults and whether or not the thinking ability can even be controlled. If not specified by the user, should there be a default or should the template be able to check if it was explicitly set? |
||
---|---|---|
.. | ||
testdata | ||
alfred.gotmpl | ||
alfred.json | ||
alpaca.gotmpl | ||
alpaca.json | ||
chatml.gotmpl | ||
chatml.json | ||
chatqa.gotmpl | ||
chatqa.json | ||
codellama-70b-instruct.gotmpl | ||
codellama-70b-instruct.json | ||
command-r.gotmpl | ||
command-r.json | ||
falcon-instruct.gotmpl | ||
falcon-instruct.json | ||
gemma-instruct.gotmpl | ||
gemma-instruct.json | ||
gemma3-instruct.gotmpl | ||
gemma3-instruct.json | ||
granite-instruct.gotmpl | ||
granite-instruct.json | ||
index.json | ||
llama2-chat.gotmpl | ||
llama2-chat.json | ||
llama3-instruct.gotmpl | ||
llama3-instruct.json | ||
magicoder.gotmpl | ||
magicoder.json | ||
mistral-instruct.gotmpl | ||
mistral-instruct.json | ||
openchat.gotmpl | ||
openchat.json | ||
phi-3.gotmpl | ||
phi-3.json | ||
solar-instruct.gotmpl | ||
solar-instruct.json | ||
starcoder2-instruct.gotmpl | ||
starcoder2-instruct.json | ||
template.go | ||
template_test.go | ||
vicuna.gotmpl | ||
vicuna.json | ||
zephyr.gotmpl | ||
zephyr.json |