hqq-quantization
Enables fast, calibration-free quantization of LLMs to 4/3/2-bit precision, optimizing memory efficiency and inference speed.
Install this skill
or
hqq-quantization3 files
Comments
Sign in to leave a comment.
No comments yet. Be the first to comment!
Install this skill with one command
/learn @fabioeducacross/hqq-quantization