hqq-quantization
Über
HQQ ermöglicht eine schnelle, kalibrierungsfreie Quantisierung von LLMs bis zu 4/3/2-Bit-Präzision ohne den Bedarf eines Datensatzes. Es ist ideal für schnelle Quantisierungs-Workflows und den Einsatz mit vLLM oder HuggingFace Transformers. Zu den Hauptvorteilen zählen eine deutlich schnellere Quantisierung als bei Methoden wie GPTQ und die Unterstützung für das Fine-Tuning quantisierter Modelle.
Schnellinstallation
Claude Code
Empfohlennpx skills add davila7/claude-code-templates -a claude-code/plugin add https://github.com/davila7/claude-code-templatesgit clone https://github.com/davila7/claude-code-templates.git ~/.claude/skills/hqq-quantizationKopieren Sie diesen Befehl und fügen Sie ihn in Claude Code ein, um diese Fähigkeit zu installieren
GitHub Repository
Verwandte Skills
quantizing-models-bitsandbytes
AndereThis skill quantizes LLMs to 8-bit or 4-bit precision using bitsandbytes, achieving 50-75% memory reduction with minimal accuracy loss. It's ideal for running larger models on limited GPU memory or accelerating inference, supporting formats like INT8, NF4, and FP4. The skill integrates with HuggingFace Transformers and enables QLoRA training and 8-bit optimizers.
gguf-quantization
DesignThis skill enables GGUF quantization for efficient model deployment on consumer hardware like CPUs and Apple Silicon. It provides flexible 2-8 bit quantization options without requiring GPU acceleration. Use it when optimizing models for local inference tools or resource-constrained environments.
awq-quantization
AndereAWQ is a 4-bit weight quantization technique that uses activation patterns to preserve critical weights, enabling 3x faster inference with minimal accuracy loss. It's ideal for deploying large models (7B-70B) on limited GPU memory and is particularly effective for instruction-tuned and multimodal models. This skill integrates with vLLM and Marlin kernels for optimized deployment.
lambda-labs-gpu-cloud
AndereThis Claude Skill provisions dedicated GPU cloud instances from Lambda Labs for ML training and inference. It's ideal for developers needing full SSH access, persistent storage, or large multi-node clusters with pre-installed stacks like PyTorch. Use it for long-running jobs where simple pricing and high-performance GPUs are required.
