Ollama

Latest version: v0.4.7

Safety actively analyzes 724004 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 15 of 19

0.1.5

What's Changed
* Fix an issue where an error would occur when running `falcon` or `starcoder` models


**Full Changelog**: https://github.com/jmorganca/ollama/compare/v0.1.4...v0.1.5

0.1.4

New models

* [OpenHermes 2 Mistral](https://ollama.ai/library/openhermes2-mistral): a new fine-tuned model based on Mistral, trained on open datasets totalling over 900,000 instructions. This model has strong multi-turn chat skills, surpassing previous Hermes 13B models and even matching 70B models on some benchmarks.

What's Changed
* Faster model switching: models will now stay loaded between requests when using different parameters (e.g. `temperature`) or system prompts
* `starcoder`, `sqlcoder` and `falcon` models now have unicode support. Note: they will need to be re-pulled (e.g. `ollama pull starcoder`)
* New documentation guide on [importing existing models](https://github.com/jmorganca/ollama/blob/main/docs/import.md) to Ollama (GGUF, PyTorch, etc)
* `ollama serve` will now print the current version of Ollama on start
* `ollama run` will now show more descriptive errors when encountering runtime issues (such as insufficient memory)
* Fixed an issue where Ollama on Linux would use CPU instead of using both the CPU and GPU for GPUs with less memory
* Fixed architecture check in Linux install script
* Fixed issue where leading whitespaces would be returned in responses
* Fixed issue where `ollama show` would show an empty `SYSTEM` prompt (instead of omitting it)
* Fixed issue with the `/api/tags` endpoint would return `null` instead of `[]` if no models were found
* Fixed an issue where `ollama show` wouldn't work when connecting remotely by using `OLLAMA_HOST`
* Fixed issue where GPU/Metal would be used on macOS even with `num_gpu` set to `0`
* Fixed issue where certain characters would be escaped in responses
* Fixed `ollama serve` logs to report the proper amount of GPU memory (VRAM) being used

Note: the `EMBED` keyword in `Modelfile` is being revisited until a future version of Ollama. Join [the discussion](https://github.com/jmorganca/ollama/issues/834) on how we can make it better.

New Contributors
* vieux made their first contribution in https://github.com/jmorganca/ollama/pull/810
* s-kostyaev made their first contribution in https://github.com/jmorganca/ollama/pull/801
* ggozad made their first contribution in https://github.com/jmorganca/ollama/pull/794
* awaescher made their first contribution in https://github.com/jmorganca/ollama/pull/811
* deichbewohner made their first contribution in https://github.com/jmorganca/ollama/pull/799

**Full Changelog**: https://github.com/jmorganca/ollama/compare/v0.1.3...v0.1.4

0.1.3

What's Changed
* Improved various API error messages to be easier to read
* Improved GPU allocation for older GPUs to fix "out of memory" errors
* Fixed issue where setting `num_gpu` to `0` would result in an error
* Ollama for macOS will now always update to the latest version, even if earlier updates had also been downloaded beforehand

**Full Changelog**: https://github.com/jmorganca/ollama/compare/v0.1.2...v0.1.3

0.1.2

New Models
* [Zephyr](https://ollama.ai/library/zephyr) A fine-tuned 7B version of mistral that was trained on a mix of publicly available, synthetic datasets and performs as well as Llama 2 70B in many benchmarks
* [Mistral OpenOrca](https://ollama.ai/library/mistral-openorca) a 7 billion parameter model fine-tuned on top of the Mistral 7B model using the OpenOrca dataset

Examples
Ollama's [examples](https://github.com/jmorganca/ollama/tree/main/examples) have been updated with some new examples:
* [Ask the mentors](https://github.com/jmorganca/ollama/tree/main/examples/typescript-mentors): a TypesScript, multi-user conversation app
* [TypeScript LangChain](https://github.com/jmorganca/ollama/tree/main/examples/langchain-typescript-simple): a simple example of using Ollama with LangChainJS and TypeScript.


What's Changed
* Download speeds for `ollama pull` have been significantly improved, from 60MB/s to over 1.5GB/s (25x faster) on fast network connections
* The API now supports non-streaming responses. Set the `stream` parameter to `false` and endpoints will return data in one single response:

curl -X POST http://localhost:11434/api/generate -d '{
"model": "llama2",
"prompt": "Why is the sky blue?",
"stream": false
}'

* Ollama can now be used with http proxies (using `HTTP_PROXY=http://<proxy>`) and https proxies (using `HTTPS_PROXY=https://<proxy>`)
* Fixed `token too long` error when generating a response
* `q8_0`, `q5_0`, `q5_1`, and `f32` models will now use GPU on Linux
* Revise help text in `ollama run` to be easier to read
* Rename runner subprocess to `ollama-runner`
* `ollama create` will now show feedback when reading model metadata
* Fix `not found error` showing when running `ollama pull`
* Improved video memory allocation on Linux to fix errors when using Nvidia GPUs

New Contributors
* xyproto made their first contribution in https://github.com/jmorganca/ollama/pull/705
* konsalex made their first contribution in https://github.com/jmorganca/ollama/pull/741

**Full Changelog**: https://github.com/jmorganca/ollama/compare/v0.1.1...v0.1.2

0.1.1

What's Changed
* Cancellable responses: `Ctrl+C` will now cancel responses when running `ollama run`
* Exit `ollama run` sessions with `Ctrl+D` or `/bye`
* Improved error messages for unknown `/slash` commands when using `ollama run`
* Various improvements to the Linux install script for distro compatibility and to fix bugs
* Fixed install issues on Fedora
* Fixed issue where specifying the `library/` prefix in `ollama run` would cause an error
* Fixed highlight color for placeholder text in `ollama run`
* Fixed issue where auto updater would not restart when clicking "Restart to Update"
* Ollama will now clean up subdirectories in `~/.ollama/models`
* Ollama when now show a default message when `ollama show` results in an empty message

New Contributors
* aaroncoffey made their first contribution in https://github.com/jmorganca/ollama/pull/629
* lstep made their first contribution in https://github.com/jmorganca/ollama/pull/621
* JayNakrani made their first contribution in https://github.com/jmorganca/ollama/pull/632
* Jimexist made their first contribution in https://github.com/jmorganca/ollama/pull/664
* hallh made their first contribution in https://github.com/jmorganca/ollama/pull/663

**Full Changelog**: https://github.com/jmorganca/ollama/compare/v0.1.0...v0.1.1

0.1.0

Ollama for Linux
<img src="https://github.com/jmorganca/ollama/assets/251292/89f8526e-866a-4e19-a73c-3ff850d45c76" height="220">

Ollama for Linux is now available, with GPU acceleration enabled out-of-the-box for Nvidia GPUs.

💯 Ollama will run on cloud servers with multiple GPUs attached
🤖 Ollama will run on WSL 2 with GPU support
😍 Ollama maximizes the number of GPU layers to load to increase performance without crashing
🤩 Ollama will support CPU only, and small hobby gaming GPUs to super powerful workstation graphics cards like the H100

Download


curl https://ollama.ai/install.sh | sh


Manual [install steps](https://github.com/jmorganca/ollama/blob/main/docs/linux.md) are also available.

Changelog
* Ollama will now automatically offload as much of the running model as is supported by your GPU for maximum performance without any crashes
* Fix issue where characters would be erased when running `ollama run`
* Added a new community project by TwanLuttik in https://github.com/jmorganca/ollama/pull/574

New Contributors
* TwanLuttik made their first contribution in https://github.com/jmorganca/ollama/pull/574

**Full Changelog**: https://github.com/jmorganca/ollama/compare/v0.0.21...v0.1.0

Page 15 of 19

© 2025 Safety CLI Cybersecurity Inc. All Rights Reserved.