Convert and quantize llm models
-
Updated
Dec 30, 2025 - Python
Convert and quantize llm models
Unified Local AI Interface & LLM Runtime (Support GGUF, Ollama, OpenAI, Gemini, etc.). Insearch of building sovereign AI system ✨
Auto GGUF Converter for HuggingFace Hub Models with Multiple Quantizations (GGUF Format)
Quantize LLMs automatically.
Create optimized GGUF quantizations by cloning from any GGUF of the same architecture.
Gemma-4-It fine-tuned on PubMedQA using SFT & RLVR
Add a description, image, and links to the gguf-quantization topic page so that developers can more easily learn about it.
To associate your repository with the gguf-quantization topic, visit your repo's landing page and select "manage topics."