Ollama

Latest version: v0.3.3

Safety actively analyzes 682244 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 6 of 15

0.1.38

New Models
- [Falcon 2](https://ollama.com/library/falcon2): A new 11B parameters causal decoder-only model built by TII and trained over 5T tokens.
- [Yi 1.5](https://ollama.com/library/yi): A new high-performing version of Yi, now licensed as Apache 2.0. Available in [6B](https://ollama.com/library/yi:6b), [9B](https://ollama.com/library/yi:9b) and [34B](https://ollama.com/library/yi:34b) sizes.

What's Changed

`ollama ps`

A new command is now available: `ollama ps`. This command displays currently loaded models, their memory footprint, and the processors used (GPU or CPU):


% ollama ps
NAME ID SIZE PROCESSOR UNTIL
mixtral:latest 7708c059a8bb 28 GB 47%/53% CPU/GPU Forever
llama3:latest a6990ed6be41 5.5 GB 100% GPU 4 minutes from now
all-minilm:latest 1b226e2802db 585 MB 100% GPU 4 minutes from now


`/clear`

To clear the chat history for a session when running `ollama run`, use `/clear`:


>>> /clear
Cleared session context


* Fixed issue where switching loaded models on Windows would take several seconds
* Running `/save` will no longer abort the chat session if an incorrect name is provided
* The `/api/tags` API endpoint will now correctly return an empty list `[]` instead of `null` if no models are provided

New Contributors
* fangtaosong made their first contribution in https://github.com/ollama/ollama/pull/4387
* machimachida made their first contribution in https://github.com/ollama/ollama/pull/4424

**Full Changelog**: https://github.com/ollama/ollama/compare/v0.1.37...v0.1.38

0.1.37

What's Changed
* Fixed issue where models with uppercase characters in the name would not show with `ollama list`
* Fixed usage string for `ollama create`
* Fix `finish_reason` being `""` instead of `null` in the Open-AI compatible chat API.

New Contributors
* todashuta made their first contribution in https://github.com/ollama/ollama/pull/4362

**Full Changelog**: https://github.com/ollama/ollama/compare/v0.1.36...v0.1.37

0.1.36

What's Changed
* Fixed `exit status 0xc0000005` error with AMD graphics cards on Windows
* Fixed rare out of memory errors when loading a model to run with CPU

**Full Changelog**: https://github.com/ollama/ollama/compare/v0.1.35...v0.1.36

0.1.35

New models
- [Llama 3 ChatQA](https://ollama.com/library/llama3-chatqa): A model from NVIDIA based on Llama 3 that excels at conversational question answering (QA) and retrieval-augmented generation (RAG).

What's Changed
* Quantization: `ollama create` can now quantize models when importing them using the `--quantize` or `-q` flag:


ollama create -f Modelfile --quantize q4_0 mymodel


> [!NOTE]
> `--quantize` works when importing `float16` or `float32` models:
> * From a binary GGUF files (e.g. `FROM ./model.gguf`)
> * From a library model (e.g. `FROM llama3:8b-instruct-fp16`)

* Fixed issue where inference subprocesses wouldn't be cleaned up on shutdown.
* Fixed a series out of memory errors when loading models on multi-GPU systems
* <kbd>Ctrl+J</kbd> characters will now properly add newlines in `ollama run`
* Fixed issues when running `ollama show` for vision models
* `OPTIONS` requests to the Ollama API will no longer result in errors
* Fixed issue where partially downloaded files wouldn't be cleaned up
* Added a new `done_reason` field in responses describing why generation stopped responding
* Ollama will now more accurately estimate how much memory is available on multi-GPU systems especially when running different models one after another

New Contributors
* fmaclen made their first contribution in https://github.com/ollama/ollama/pull/3884
* Renset made their first contribution in https://github.com/ollama/ollama/pull/3881
* glumia made their first contribution in https://github.com/ollama/ollama/pull/3043
* boessu made their first contribution in https://github.com/ollama/ollama/pull/4236
* gaardhus made their first contribution in https://github.com/ollama/ollama/pull/2307
* svilupp made their first contribution in https://github.com/ollama/ollama/pull/2192
* WolfTheDeveloper made their first contribution in https://github.com/ollama/ollama/pull/4300

**Full Changelog**: https://github.com/ollama/ollama/compare/v0.1.34...v0.1.35

0.1.34

![Ollama goes on an adventure to hunt down bugs](https://github.com/ollama/ollama/assets/3325447/e878f0d0-5a88-448c-a88c-53146061ae3b)

New models
* [Llava Llama 3](https://ollama.com/library/llava-llama3): A new high-performing LLaVA model fine-tuned from Llama 3 Instruct.
* [Llava Phi 3](https://ollama.com/library/llava-phi3): A new small LLaVA model fine-tuned from Phi 3.
* [StarCoder2 15B Instruct](https://ollama.com/library/starcoder2:15b-instruct): A new instruct fine-tune of the StarCoder2 model
* [CodeGemma 1.1](https://ollama.com/library/codegemma): A new release of the CodeGemma model.
* [StableLM2 12B](https://ollama.com/library/stablelm2:12b): A new 12B version of the StableLM 2 model from Stability AI
* [Moondream 2](https://ollama.com/library/moondream): Moondream 2's runtime parameters have been improved for better responses

What's Changed
* Fixed issues with LLaVa models where they would respond incorrectly after the first request
* Fixed out of memory errors when running large models such as Llama 3 70B
* Fixed various issues with Nvidia GPU discovery on Linux and Windows
* Fixed a series of Modelfile errors when running `ollama create`
* Fixed `no slots available` error that occurred when cancelling a request and then sending follow up requests
* Improved AMD GPU detection on Fedora
* Improved reliability when using the experimental `OLLAMA_NUM_PARALLEL` and `OLLAMA_MAX_LOADED_MODELS` flags
* `ollama serve` will now shut down quickly, even if a model is loading

New Contributors
* drnic made their first contribution in https://github.com/ollama/ollama/pull/4116
* bernardo-bruning made their first contribution in https://github.com/ollama/ollama/pull/4111
* Drlordbasil made their first contribution in https://github.com/ollama/ollama/pull/4174
* Saif-Shines made their first contribution in https://github.com/ollama/ollama/pull/4119
* HydenLiu made their first contribution in https://github.com/ollama/ollama/pull/4194
* jl-codes made their first contribution in https://github.com/ollama/ollama/pull/3621
* Nurgo made their first contribution in https://github.com/ollama/ollama/pull/3473
* adrienbrault made their first contribution in https://github.com/ollama/ollama/pull/3129
* Darinochka made their first contribution in https://github.com/ollama/ollama/pull/3945

**Full Changelog**: https://github.com/ollama/ollama/compare/v0.1.33...v0.1.34

0.1.33

![Llama 3](https://github.com/ollama/ollama/assets/3325447/8dc9c472-9d72-4b39-95ae-2c85ada375b9)

New models:
* [Llama 3](https://ollama.com/library/llama3): a new model by Meta, and the most capable openly available LLM to date
* [Phi 3 Mini](https://ollama.com/library/phi3): a new 3.8B parameters, lightweight, state-of-the-art open model by Microsoft.
* [Moondream](https://ollama.com/library/moondream) moondream is a small vision language model designed to run efficiently on edge devices.
* [Llama 3 Gradient 1048K](https://ollama.com/library/llama3-gradient): A Llama 3 fine-tune by Gradient to support up to a 1M token context window.
* [Dolphin Llama 3](https://ollama.com/library/dolphin-llama3): The uncensored Dolphin model, trained by Eric Hartford and based on Llama 3 with a variety of instruction, conversational, and coding skills.
* [Qwen 110B](https://ollama.com/library/qwen:110b): The first Qwen model over 100B parameters in size with outstanding performance in evaluations

What's Changed
* Fixed issues where the model would not terminate, causing the API to hang.
* Fixed a series of out of memory errors on Apple Silicon Macs
* Fixed out of memory errors when running Mixtral architecture models

Experimental concurrency features

New concurrency features are coming soon to Ollama. They are available

* `OLLAMA_NUM_PARALLEL`: Handle multiple requests simultaneously for a single model
* `OLLAMA_MAX_LOADED_MODELS`: Load multiple models simultaneously

To enable these features, set the environment variables for `ollama serve`. For more info see [this guide](https://github.com/ollama/ollama/blob/main/docs/faq.md#how-do-i-configure-ollama-server):


OLLAMA_NUM_PARALLEL=4 OLLAMA_MAX_LOADED_MODELS=4 ollama serve


New Contributors
* hmartinez82 made their first contribution in https://github.com/ollama/ollama/pull/3972
* Cephra made their first contribution in https://github.com/ollama/ollama/pull/4037
* arpitjain099 made their first contribution in https://github.com/ollama/ollama/pull/4007
* MarkWard0110 made their first contribution in https://github.com/ollama/ollama/pull/4031
* alwqx made their first contribution in https://github.com/ollama/ollama/pull/4073
* sidxt made their first contribution in https://github.com/ollama/ollama/pull/3705
* ChengenH made their first contribution in https://github.com/ollama/ollama/pull/3789
* secondtruth made their first contribution in https://github.com/ollama/ollama/pull/3503
* reid41 made their first contribution in https://github.com/ollama/ollama/pull/3612
* ericcurtin made their first contribution in https://github.com/ollama/ollama/pull/3626
* JT2M0L3Y made their first contribution in https://github.com/ollama/ollama/pull/3633
* datvodinh made their first contribution in https://github.com/ollama/ollama/pull/3655
* MapleEve made their first contribution in https://github.com/ollama/ollama/pull/3817
* swuecho made their first contribution in https://github.com/ollama/ollama/pull/3810
* brycereitano made their first contribution in https://github.com/ollama/ollama/pull/3895
* bsdnet made their first contribution in https://github.com/ollama/ollama/pull/3889
* fyxtro made their first contribution in https://github.com/ollama/ollama/pull/3855
* natalyjazzviolin made their first contribution in https://github.com/ollama/ollama/pull/3962

**Full Changelog**: https://github.com/ollama/ollama/compare/v0.1.32...v0.1.33

Page 6 of 15

© 2024 Safety CLI Cybersecurity Inc. All Rights Reserved.