Llama-cpp-python

Latest version: v0.3.8

Safety actively analyzes 723576 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 20 of 22

0.1.71

Not secure
- (llama.cpp) Update llama.cpp

- (server) Fix several pydantic v2 migration bugs

0.1.70

Not secure
- (Llama.create_completion) Revert change so that `max_tokens` is not truncated to `context_size` in `create_completion`
- (server) Fixed changed settings field names from pydantic v2 migration

0.1.69

Not secure
- (server) Streaming requests can are now interrupted pre-maturely when a concurrent request is made. Can be controlled with the `interrupt_requests` setting.
- (server) Moved to fastapi v0.100.0 and pydantic v2
- (docker) Added a new "simple" image that builds llama.cpp from source when started.
- (server) performance improvements by avoiding unnecessary memory allocations during sampling

0.1.68

Not secure
- (llama.cpp) Update llama.cpp

0.1.67

Not secure
- Fix performance bug in Llama model by pre-allocating memory tokens and logits.
- Fix bug in Llama model where the model was not free'd after use.

0.1.66

Not secure
- (llama.cpp) New model API

- Performance issue during eval caused by looped np.concatenate call
- State pickling issue when saving cache to disk

Page 20 of 22

© 2025 Safety CLI Cybersecurity Inc. All Rights Reserved.