Llama-cpp-cffi

Latest version: v0.4.40

Safety actively analyzes 715170 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 13 of 13

0.1.4

Added:
- README: supported GPU Compute Capability for CUDA.

Fixed:
- Cleaned up `build.py`.
- Type annotations in OpenAI related code.

0.1.3

Added:
- Support for PyPy 3.10 versions.

Changed:
- Disabled GitHub Actions.
- Uses `upx -9` to compress shared/dynamic libraries.
- Repacks `whl` with better compression rate.
- Auto-detect CUDA support.

Removed:
- ctypes version and demos.

0.1.2

Added:
- Preparation for [Chat Completions API by OpenAI ©](https://platform.openai.com/docs/overview) compatible server.

Fixed:
- Argument `options` is `deepcopy`-ed when passed to `llama_generate(options)`, so it can be reused.

Changed:
- Build for `manylinux_2_28` and `musllinux_1_2`.
- Build for [CUDA Compute Capability](https://developer.nvidia.com/cuda-gpus) >= 6.1.

0.1.1

Changed:
- Updated: `huggingface-hub = "^0.24.0"`, `setuptools = "^71.0.3"`

0.1.0

Added:
- Park first version.

Page 13 of 13

© 2025 Safety CLI Cybersecurity Inc. All Rights Reserved.