Latest version: v2.2.0
The information on this page was curated by experts in our Cybersecurity Intelligence Team.
Production ready LLM model compression/quantization toolkit with hw accelerated inference support for both cpu/gpu via HF, vLLM, and SGLang.
No known vulnerabilities found
Has known vulnerabilities