New models
- [Cohere Aya 23](https://ollama.com/library/aya): A new state-of-the-art, multilingual LLM covering 23 different languages.
- [Mistral 7B 0.3](https://ollama.com/library/mistral:v0.3): A new version of Mistral 7B with initial support for function calling.
- [Phi-3 Medium](https://ollama.com/library/phi3:medium): a 14B parameters, lightweight, state-of-the-art open model by Microsoft.
- [Phi-3 Mini 128K](https://ollama.com/library/phi3:mini-128k) and [Phi-3 Medium 128K](https://ollama.com/library/phi3:medium-128k): versions of the Phi-3 models that support a context window size of 128K
- [Granite code](https://ollama.com/library/granite-code): A family of open foundation models by IBM for Code Intelligence
Llama 3 import
It is now possible to import and quantize Llama 3 and its finetunes from Safetensors format to Ollama.
First, clone a Hugging Face repo with a Safetensors model:
git clone https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct
cd Meta-Llama-3-8B-Instruct
Next, create a `Modelfile`:
FROM .
TEMPLATE """{{ if .System }}<|start_header_id|>system<|end_header_id|>
{{ .System }}<|eot_id|>{{ end }}{{ if .Prompt }}<|start_header_id|>user<|end_header_id|>
{{ .Prompt }}<|eot_id|>{{ end }}<|start_header_id|>assistant<|end_header_id|>
{{ .Response }}<|eot_id|>"""
PARAMETER stop <|start_header_id|>
PARAMETER stop <|end_header_id|>
PARAMETER stop <|eot_id|>
Then, create and quantize a model:
ollama create --quantize q4_0 -f Modelfile my-llama3
ollama run my-llama3
What's Changed
* Fixed issues where wide characters such as Chinese, Korean, Japanese and Russian languages.
* Added new `OLLAMA_NOHISTORY=1` environment variable that can be set to disable history when using `ollama run`
* New experimental `OLLAMA_FLASH_ATTENTION=1` flag for `ollama serve` that improves token generation speed on Apple Silicon Macs and NVIDIA graphics cards
* Fixed error that would occur on Windows running `ollama create -f Modelfile`
* `ollama create` can now create models from I-Quant GGUF files
* Fixed `EOF` errors when resuming downloads via `ollama pull`
* Added a `Ctrl+W` shortcut to `ollama run`
New Contributors
* rapmd73 made their first contribution in https://github.com/ollama/ollama/pull/4467
* sammcj made their first contribution in https://github.com/ollama/ollama/pull/4120
* likejazz made their first contribution in https://github.com/ollama/ollama/pull/4535
**Full Changelog**: https://github.com/ollama/ollama/compare/v0.1.38...v0.1.39