MCP HubMCP Hub
Zurück zu Fähigkeiten

model-selection

vamseeachanta
Aktualisiert Today
62 Ansichten
3
2
3
Auf GitHub ansehen
Andere

Über

Diese Fähigkeit bietet automatisierte Modellauswahlberatung für Claude Code basierend auf Aufgabenkomplexität, Kosten und Latenzanforderungen. Sie stellt einen Entscheidungsbaum und eine Kurzreferenztabelle bereit, um Entwicklern bei der effizienten Auswahl zwischen Opus-, Sonnet- und Haiku-Modellen zu helfen. Nutzen Sie sie bei neuen Aufgaben, um Leistung und Ressourcennutzung zu optimieren.

Schnellinstallation

Claude Code

Empfohlen
Primär
npx skills add vamseeachanta/workspace-hub
Plugin-BefehlAlternativ
/plugin add https://github.com/vamseeachanta/workspace-hub
Git CloneAlternativ
git clone https://github.com/vamseeachanta/workspace-hub.git ~/.claude/skills/model-selection

Kopieren Sie diesen Befehl und fügen Sie ihn in Claude Code ein, um diese Fähigkeit zu installieren

Dokumentation

Model Selection Skill

Version: 1.0.0 Category: Optimization Triggers: Starting tasks, choosing Claude model, usage optimization

Quick Reference

Model Selection Decision Tree

NEW TASK
    │
    ├── WORK REPO + COMPLEX → OPUS
    ├── WORK REPO + STANDARD → SONNET
    ├── PERSONAL + SIMPLE → HAIKU
    └── DEFAULT → SONNET

Quick Selection Guide

ModelTarget %Use For
OPUS30%Architecture, multi-file refactoring (>5 files), security review
SONNET40%Standard implementations, code review, documentation
HAIKU30%Quick queries, status checks, simple operations

Automated Model Suggestion

# Get model recommendation before each task
./scripts/monitoring/suggest_model.sh <repository> "<task description>"

# Examples:
./scripts/monitoring/suggest_model.sh digitalmodel "Design authentication architecture"
# → Recommends: OPUS (complexity score: 4)

./scripts/monitoring/suggest_model.sh digitalmodel "Implement user login"
# → Recommends: SONNET (complexity score: 1)

./scripts/monitoring/suggest_model.sh hobbies "Quick file check"
# → Recommends: HAIKU (complexity score: -3)

Complexity Scoring

Algorithm evaluates:

  1. Keywords - architecture/refactor → +3, implement/feature → +1, check/status → -2
  2. Repository Tier - Work Tier 1 → +1, Personal → -1
  3. Task Length - >15 words → +1, <5 words → -1

Score Mapping:

  • Score ≥3: OPUS
  • Score 0-2: SONNET
  • Score <0: HAIKU

Repository Tiers

Work Repositories

Tier 1 (Production): 60% Opus, 30% Sonnet, 10% Haiku

  • digitalmodel, energy, frontierdeepwater

Tier 2 (Active): 30% Opus, 50% Sonnet, 20% Haiku

  • assetutilities, worldenergydata

Tier 3 (Maintenance): 10% Opus, 30% Sonnet, 60% Haiku

  • doris, saipem, OGManufacturing

Personal Repositories

Active: 20% Opus, 40% Sonnet, 40% Haiku Experimental: 5% Opus, 25% Sonnet, 70% Haiku Archive: 0% Opus, 20% Sonnet, 80% Haiku

Usage Monitoring

Check before starting work: https://claude.ai/settings/usage

Alert Thresholds:

  • Sonnet >70% → Switch to Opus/Haiku
  • Session >80% → Batch work or wait
  • Overall >80% → Defer non-critical

OPUS Use Cases

✅ Multi-file refactoring (>5 files) ✅ Architecture decisions ✅ Complex algorithm design ✅ Security-critical code review ✅ Cross-repository coordination ✅ Performance optimization strategies

SONNET Use Cases

✅ Standard feature implementation ✅ Code review (single PR) ✅ Documentation writing ✅ Test generation ✅ Bug fixing (standard complexity) ✅ Configuration updates

HAIKU Use Cases

✅ File existence checks ✅ Simple grep/search operations ✅ Quick status updates ✅ Log analysis (pattern matching) ✅ Template generation ✅ Format validation

Emergency Protocols

If Sonnet >80%

⛔ STOP using Sonnet immediately
✅ Switch to Opus for critical work
✅ Switch to Haiku for everything else
📅 Defer non-urgent work to Tuesday

If Session >80%

⏸️  Pause AI tasks
⏰ Wait for session reset (~3-4 hours)
📦 Batch work for next session

Full Reference

See: @docs/AI_MODEL_SELECTION_AUTOMATION.md See: @docs/CLAUDE_MODEL_SELECTION_QUICK_REFERENCE.md


Use this when starting tasks, selecting models, or optimizing AI usage.

GitHub Repository

vamseeachanta/workspace-hub
Pfad: .claude/skills/ai/optimization/model-selection

Verwandte Skills

data-science-expert

Andere

This skill provides expert-level data science assistance including statistical analysis, machine learning, and data visualization. It helps developers with tasks like data cleaning, model building, and creating plots using Python libraries like pandas and matplotlib. Use it when you need guidance on EDA, statistical modeling, or visualizing complex datasets.

Skill ansehen

usage-optimization

Andere

This Claude Skill helps developers optimize AI usage efficiency by promoting script-first patterns, batch operations, and prepared input files. It provides effectiveness ratings for different approaches, emphasizing execution over description to maximize productivity. Use it when you need to reduce AI interaction time and get more actionable, automated outputs from Claude.

Skill ansehen

agent-usage-optimizer

Andere

This skill optimizes AI model selection by reading quota states and recommending the best Claude/Codex/Gemini allocation for each task. It provides quota-aware routing and headroom displays, making it ideal for work sessions with multiple queued items or when approaching quota limits. Developers should use it before starting sessions with 3+ work items or when Claude quotas drop below 50% remaining.

Skill ansehen

agent-usage-optimizer-provider-capability-reference

Andere

This reference skill provides a quick comparison table of AI providers (Claude models, Codex, Gemini) with their strengths and ideal use cases. It helps developers optimize agent usage by selecting the most suitable provider for tasks like architecture, coding, or bulk data processing. Use it to make informed decisions on model selection based on task requirements and constraints like quota and cost.

Skill ansehen