Latest version: v2.11.0
The information on this page was curated by experts in our Cybersecurity Intelligence Team.
AirLLM allows single 4GB GPU card to run 70B large language models without quantization, distillation or pruning. 8GB vmem to run 405B Llama3.1.
No known vulnerabilities found
Has known vulnerabilities