mcp-management
About
This skill enables developers to manage and interact with Model Context Protocol (MCP) servers, allowing you to discover, analyze, and execute tools, prompts, and resources from configured MCP servers. It supports intelligent tool selection, multi-server management, and context-efficient capability discovery without polluting the main context window. Use it when implementing MCP client functionality or needing to programmatically access and filter MCP capabilities for specific tasks.
Documentation
MCP Management
Skill for managing and interacting with Model Context Protocol (MCP) servers.
Overview
MCP is an open protocol enabling AI agents to connect to external tools and data sources. This skill provides scripts and utilities to discover, analyze, and execute MCP capabilities from configured servers without polluting the main context window.
Key Benefits:
- Progressive disclosure of MCP capabilities (load only what's needed)
- Intelligent tool/prompt/resource selection based on task requirements
- Multi-server management from single config file
- Context-efficient: subagents handle MCP discovery and execution
- Persistent tool catalog: automatically saves discovered tools to JSON for fast reference
When to Use This Skill
Use this skill when:
- Discovering MCP Capabilities: Need to list available tools/prompts/resources from configured servers
- Task-Based Tool Selection: Analyzing which MCP tools are relevant for a specific task
- Executing MCP Tools: Calling MCP tools programmatically with proper parameter handling
- MCP Integration: Building or debugging MCP client implementations
- Context Management: Avoiding context pollution by delegating MCP operations to subagents
Core Capabilities
1. Configuration Management
MCP servers configured in .claude/.mcp.json.
Gemini CLI Integration (recommended): Create symlink to .gemini/settings.json:
mkdir -p .gemini && ln -sf .claude/.mcp.json .gemini/settings.json
See references/configuration.md and references/gemini-cli-integration.md.
2. Capability Discovery
npx tsx scripts/cli.ts list-tools # Saves to assets/tools.json
npx tsx scripts/cli.ts list-prompts
npx tsx scripts/cli.ts list-resources
Aggregates capabilities from multiple servers with server identification.
3. Intelligent Tool Analysis
LLM analyzes assets/tools.json directly - better than keyword matching algorithms.
4. Tool Execution
Primary: Gemini CLI (if available)
gemini -y -m gemini-2.5-flash -p "Take a screenshot of https://example.com"
Secondary: Direct Scripts
npx tsx scripts/cli.ts call-tool memory create_entities '{"entities":[...]}'
Fallback: mcp-manager Subagent
See references/gemini-cli-integration.md for complete examples.
Implementation Patterns
Pattern 1: Gemini CLI Auto-Execution (Primary)
Use Gemini CLI for automatic tool discovery and execution. See references/gemini-cli-integration.md for complete guide.
Quick Example:
gemini -y -m gemini-2.5-flash -p "Take a screenshot of https://example.com"
Benefits: Automatic tool discovery, natural language execution, faster than subagent orchestration.
Pattern 2: Subagent-Based Execution (Fallback)
Use mcp-manager agent when Gemini CLI unavailable. Subagent discovers tools, selects relevant ones, executes tasks, reports back.
Benefit: Main context stays clean, only relevant tool definitions loaded when needed.
Pattern 3: LLM-Driven Tool Selection
LLM reads assets/tools.json, intelligently selects relevant tools using context understanding, synonyms, and intent recognition.
Pattern 4: Multi-Server Orchestration
Coordinate tools across multiple servers. Each tool knows its source server for proper routing.
Scripts Reference
scripts/mcp-client.ts
Core MCP client manager class. Handles:
- Config loading from
.claude/.mcp.json - Connecting to multiple MCP servers
- Listing tools/prompts/resources across all servers
- Executing tools with proper error handling
- Connection lifecycle management
scripts/cli.ts
Command-line interface for MCP operations. Commands:
list-tools- Display all tools and save toassets/tools.jsonlist-prompts- Display all promptslist-resources- Display all resourcescall-tool <server> <tool> <json>- Execute a tool
Note: list-tools persists complete tool catalog to assets/tools.json with full schemas for fast reference, offline browsing, and version control.
Quick Start
Method 1: Gemini CLI (recommended)
npm install -g gemini-cli
mkdir -p .gemini && ln -sf .claude/.mcp.json .gemini/settings.json
gemini -y -m gemini-2.5-flash -p "Take a screenshot of https://example.com"
Method 2: Scripts
cd .claude/skills/mcp-management/scripts && npm install
npx tsx cli.ts list-tools # Saves to assets/tools.json
npx tsx cli.ts call-tool memory create_entities '{"entities":[...]}'
Method 3: mcp-manager Subagent
See references/gemini-cli-integration.md for complete guide.
Technical Details
See references/mcp-protocol.md for:
- JSON-RPC protocol details
- Message types and formats
- Error codes and handling
- Transport mechanisms (stdio, HTTP+SSE)
- Best practices
Integration Strategy
Execution Priority
-
Gemini CLI (Primary): Fast, automatic, intelligent tool selection
- Check:
command -v gemini - Execute:
gemini -y -m gemini-2.5-flash -p "<task>" - Best for: All tasks when available
- Check:
-
Direct CLI Scripts (Secondary): Manual tool specification
- Use when: Need specific tool/server control
- Execute:
npx tsx scripts/cli.ts call-tool <server> <tool> <args>
-
mcp-manager Subagent (Fallback): Context-efficient delegation
- Use when: Gemini unavailable or failed
- Keeps main context clean
Integration with Agents
The mcp-manager agent uses this skill to:
- Check Gemini CLI availability first
- Execute via
geminicommand if available - Fallback to direct script execution
- Discover MCP capabilities without loading into main context
- Report results back to main agent
This keeps main agent context clean and enables efficient MCP integration.
Quick Install
/plugin add https://github.com/mrgoonie/claudekit-skills/tree/main/mcp-managementCopy and paste this command in Claude Code to install this skill
GitHub 仓库
Related Skills
sglang
MetaSGLang is a high-performance LLM serving framework that specializes in fast, structured generation for JSON, regex, and agentic workflows using its RadixAttention prefix caching. It delivers significantly faster inference, especially for tasks with repeated prefixes, making it ideal for complex, structured outputs and multi-turn conversations. Choose SGLang over alternatives like vLLM when you need constrained decoding or are building applications with extensive prefix sharing.
evaluating-llms-harness
TestingThis Claude Skill runs the lm-evaluation-harness to benchmark LLMs across 60+ standardized academic tasks like MMLU and GSM8K. It's designed for developers to compare model quality, track training progress, or report academic results. The tool supports various backends including HuggingFace and vLLM models.
llamaguard
OtherLlamaGuard is Meta's 7-8B parameter model for moderating LLM inputs and outputs across six safety categories like violence and hate speech. It offers 94-95% accuracy and can be deployed using vLLM, Hugging Face, or Amazon SageMaker. Use this skill to easily integrate content filtering and safety guardrails into your AI applications.
langchain
MetaLangChain is a framework for building LLM applications using agents, chains, and RAG pipelines. It supports multiple LLM providers, offers 500+ integrations, and includes features like tool calling and memory management. Use it for rapid prototyping and deploying production systems like chatbots, autonomous agents, and question-answering services.
