running-mutation-tests
について
このClaudeスキルは、専門的なテストシナリオで使用される、テストスイートの有効性を評価するためのミューテーションテストを実行します。コード変更を導入することで、テストがバグをどの程度効果的に検出できるかを検証するのに役立ちます。このスキルは、設定された環境内でミューテーションテストを実行するために、特定のbashコマンドを使用します。
クイックインストール
Claude Code
推奨/plugin add https://github.com/jeremylongshore/claude-code-plugins-plusgit clone https://github.com/jeremylongshore/claude-code-plugins-plus.git ~/.claude/skills/running-mutation-testsこのコマンドをClaude Codeにコピー&ペーストしてスキルをインストールします
ドキュメント
Prerequisites
Before using this skill, ensure you have:
- Test environment configured and accessible
- Required testing tools and frameworks installed
- Test data and fixtures prepared
- Appropriate permissions for test execution
- Network connectivity if testing external services
Instructions
Step 1: Prepare Test Environment
Set up the testing context:
- Use Read tool to examine configuration from {baseDir}/config/
- Validate test prerequisites are met
- Initialize test framework and load dependencies
- Configure test parameters and thresholds
Step 2: Execute Tests
Run the test suite:
- Use Bash(test:mutation-*) to invoke test framework
- Monitor test execution progress
- Capture test outputs and metrics
- Handle test failures and error conditions
Step 3: Analyze Results
Process test outcomes:
- Identify passed and failed tests
- Calculate success rate and performance metrics
- Detect patterns in failures
- Generate insights for improvement
Step 4: Generate Report
Document findings in {baseDir}/test-reports/:
- Test execution summary
- Detailed failure analysis
- Performance benchmarks
- Recommendations for fixes
Output
The skill generates comprehensive test results:
Test Summary
- Total tests executed
- Pass/fail counts and percentage
- Execution time metrics
- Resource utilization stats
Detailed Results
Each test includes:
- Test name and identifier
- Execution status (pass/fail/skip)
- Actual vs. expected outcomes
- Error messages and stack traces
Metrics and Analysis
- Code coverage percentages
- Performance benchmarks
- Trend analysis across runs
- Quality gate compliance status
Error Handling
Common issues and solutions:
Environment Setup Failures
- Error: Test environment not properly configured
- Solution: Verify configuration files; check environment variables; ensure dependencies are installed
Test Execution Timeouts
- Error: Tests exceeded maximum execution time
- Solution: Increase timeout thresholds; optimize slow tests; parallelize test execution
Resource Exhaustion
- Error: Insufficient memory or disk space during testing
- Solution: Clean up temporary files; reduce concurrent test workers; increase resource allocation
Dependency Issues
- Error: Required services or databases unavailable
- Solution: Verify service health; check network connectivity; use mocks if services are down
Resources
Testing Tools
- Industry-standard testing frameworks for your language/platform
- CI/CD integration guides and plugins
- Test automation best practices documentation
Best Practices
- Maintain test isolation and independence
- Use meaningful test names and descriptions
- Keep tests fast and focused
- Implement proper setup and teardown
- Version control test artifacts
- Run tests in CI/CD pipelines
GitHub リポジトリ
関連スキル
content-collections
メタThis skill provides a production-tested setup for Content Collections, a TypeScript-first tool that transforms Markdown/MDX files into type-safe data collections with Zod validation. Use it when building blogs, documentation sites, or content-heavy Vite + React applications to ensure type safety and automatic content validation. It covers everything from Vite plugin configuration and MDX compilation to deployment optimization and schema validation.
creating-opencode-plugins
メタThis skill provides the structure and API specifications for creating OpenCode plugins that hook into 25+ event types like commands, files, and LSP operations. It offers implementation patterns for JavaScript/TypeScript modules that intercept and extend the AI assistant's lifecycle. Use it when you need to build event-driven plugins for monitoring, custom handling, or extending OpenCode's capabilities.
evaluating-llms-harness
テストThis Claude Skill runs the lm-evaluation-harness to benchmark LLMs across 60+ standardized academic tasks like MMLU and GSM8K. It's designed for developers to compare model quality, track training progress, or report academic results. The tool supports various backends including HuggingFace and vLLM models.
polymarket
メタThis skill enables developers to build applications with the Polymarket prediction markets platform, including API integration for trading and market data. It also provides real-time data streaming via WebSocket to monitor live trades and market activity. Use it for implementing trading strategies or creating tools that process live market updates.
