exllamav3
oobabooga/exllamav3
An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs
2stars
Forks
1
Open issues
0
Watchers
2
Size
5.0 MB
PythonMIT License
Created: Apr 7, 2025
Updated: Oct 4, 2025
Last push: Aug 17, 2025
Fork