Zurück zu Fähigkeiten

training-llms-megatron

davila7
Aktualisiert 15 days ago
294 Ansichten
18,478
1,685
18,478
Auf GitHub ansehen
DesignMegatron-CoreLarge-Scale TrainingNVIDIATensor ParallelismPipeline ParallelismModel ParallelismH100Distributed TrainingProduction

Über

Diese Fähigkeit trainiert massive LLMs (2B-462B Parameter) mit NVIDIAs Megatron-Core-Framework für maximale GPU-Effizienz. Nutzen Sie sie, wenn Sie Modelle mit über 1B Parametern trainieren und erweiterte Parallelisierungsmethoden wie Tensor-, Pipeline- oder Expert-Parallelismus benötigen. Es handelt sich um ein produktionsreifes Framework, das sich bereits bei Modellen wie Nemotron und LLaMA bewährt hat.

Schnellinstallation

Claude Code

Empfohlen
Primär
npx skills add davila7/claude-code-templates -a claude-code
Plugin-BefehlAlternativ
/plugin add https://github.com/davila7/claude-code-templates
Git CloneAlternativ
git clone https://github.com/davila7/claude-code-templates.git ~/.claude/skills/training-llms-megatron

Kopieren Sie diesen Befehl und fügen Sie ihn in Claude Code ein, um diese Fähigkeit zu installieren

GitHub Repository

davila7/claude-code-templates
Pfad: cli-tool/components/skills/ai-research/distributed-training-megatron-core
0
anthropicanthropic-claudeclaudeclaude-code

Verwandte Skills

openrlhf-training

Design

OpenRLHF is a high-performance RLHF training framework for fine-tuning large language models (7B-70B+ parameters) using methods like PPO, DPO, and GRPO. It leverages Ray for distributed architecture and vLLM for accelerated inference, achieving speeds 2x faster than alternatives like DeepSpeedChat. Use this skill when you need efficient, distributed RLHF training with optimized GPU resource sharing and ZeRO-3 support.

Skill ansehen

huggingface-tokenizers

Dokumente

This skill provides high-performance tokenization using HuggingFace's Rust-based library, processing 1GB of text in under 20 seconds. It supports BPE, WordPiece, and Unigram algorithms while enabling custom tokenizer training and alignment tracking. Use it when you need production-fast tokenization or to build custom tokenizers integrated with the transformers ecosystem.

Skill ansehen

qdrant-vector-search

Meta

The qdrant-vector-search skill provides a high-performance vector similarity search engine for building production RAG systems. It enables fast nearest neighbor search, hybrid search with filtering, and scalable vector storage powered by Rust. Use it when you need low-latency semantic search with horizontal scaling capabilities and full data control.

Skill ansehen

crewai-multi-agent

Meta

CrewAI is a lightweight multi-agent orchestration framework for building teams of specialized AI agents that collaborate autonomously on complex tasks. It enables role-based agent collaboration with memory and supports sequential or hierarchical workflows for production use. The framework is built without LangChain dependencies for lean, fast execution.

Skill ansehen