Tags: #compression
LLM Optimization Toolkit
huggingface
1.1k
ModelCloud/GPTQModel
A toolkit for quantizing (compressing) Large Language Models (LLMs) with hardware acceleration across various GPUs and CPUs, integrating with popular inference frameworks.
C Library, Command-line Utility Suite
c
3.5k
libarchive/libarchive
A portable C library and suite of command-line tools for reading, writing, and manipulating a wide array of archive and compression formats.