Pinned Loading
-
ModelCloud/GPTQModel
ModelCloud/GPTQModel PublicLLM model quantization (compression) toolkit with HW acceleration support for Nvidia, AMD, Intel GPU and Intel/AMD/Apple CPU via HF, vLLM, and SGLang.
-
IST-DASLab/MoE-Quant
IST-DASLab/MoE-Quant PublicCode for data-aware compression of DeepSeek models
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.

