MCP HubMCP Hub
Volver a habilidades

code-review-quality

proffesor-for-testing
Actualizado Today
243 vistas
267
56
267
Ver en GitHub
Otrocode-reviewfeedbackqualitytestabilitymaintainabilitypr-review

Acerca de

Esta habilidad realiza revisiones automatizadas de código centradas en la calidad, la capacidad de prueba y la mantenibilidad, priorizando comentarios críticos como errores y problemas de seguridad. Está diseñada para su uso durante revisiones de código, al proporcionar retroalimentación o al establecer prácticas de revisión. La herramienta categoriza los comentarios por gravedad y enfatiza hacer preguntas en lugar de emitir órdenes.

Instalación rápida

Claude Code

Recomendado
Principal
npx skills add proffesor-for-testing/agentic-qe
Comando PluginAlternativo
/plugin add https://github.com/proffesor-for-testing/agentic-qe
Git CloneAlternativo
git clone https://github.com/proffesor-for-testing/agentic-qe.git ~/.claude/skills/code-review-quality

Copia y pega este comando en Claude Code para instalar esta habilidad

Documentación

Code Review Quality

<default_to_action> When reviewing code or establishing review practices:

  1. PRIORITIZE feedback: 🔴 Blocker (must fix) → 🟡 Major → 🟢 Minor → 💡 Suggestion
  2. FOCUS on: Bugs, security, testability, maintainability (not style preferences)
  3. ASK questions over commands: "Have you considered...?" > "Change this to..."
  4. PROVIDE context: Why this matters, not just what to change
  5. LIMIT scope: Review < 400 lines at a time for effectiveness

Quick Review Checklist:

  • Logic: Does it work correctly? Edge cases handled?
  • Security: Input validation? Auth checks? Injection risks?
  • Testability: Can this be tested? Is it tested?
  • Maintainability: Clear naming? Single responsibility? DRY?
  • Performance: O(n²) loops? N+1 queries? Memory leaks?

Critical Success Factors:

  • Review the code, not the person
  • Catching bugs > nitpicking style
  • Fast feedback (< 24h) > thorough feedback </default_to_action>

Quick Reference Card

When to Use

  • PR code reviews
  • Pair programming feedback
  • Establishing team review standards
  • Mentoring developers

Feedback Priority Levels

LevelIconMeaningAction
Blocker🔴Bug/security/crashMust fix before merge
Major🟡Logic issue/test gapShould fix before merge
Minor🟢Style/namingNice to fix
Suggestion💡Alternative approachConsider for future

Review Scope Limits

Lines ChangedRecommendation
< 200Single review session
200-400Review in chunks
> 400Request PR split

What to Focus On

✅ Review❌ Skip
Logic correctnessFormatting (use linter)
Security risksNaming preferences
Test coverageArchitecture debates
Performance issuesStyle opinions
Error handlingTrivial changes

Feedback Templates

Blocker (Must Fix)

🔴 **BLOCKER: SQL Injection Risk**

This query is vulnerable to SQL injection:
```javascript
db.query(`SELECT * FROM users WHERE id = ${userId}`)

Fix: Use parameterized queries:

db.query('SELECT * FROM users WHERE id = ?', [userId])

Why: User input directly in SQL allows attackers to execute arbitrary queries.


### Major (Should Fix)
```markdown
🟡 **MAJOR: Missing Error Handling**

What happens if `fetchUser()` throws? The error bubbles up unhandled.

**Suggestion:** Add try/catch with appropriate error response:
```javascript
try {
  const user = await fetchUser(id);
  return user;
} catch (error) {
  logger.error('Failed to fetch user', { id, error });
  throw new NotFoundError('User not found');
}

### Minor (Nice to Fix)
```markdown
🟢 **minor:** Variable name could be clearer

`d` doesn't convey meaning. Consider `daysSinceLastLogin`.

Suggestion (Consider)

💡 **suggestion:** Consider extracting this to a helper

This validation logic appears in 3 places. A `validateEmail()` helper would reduce duplication. Not blocking, but might be worth a follow-up PR.

Review Questions to Ask

Logic

  • What happens when X is null/empty/negative?
  • Is there a race condition here?
  • What if the API call fails?

Security

  • Is user input validated/sanitized?
  • Are auth checks in place?
  • Any secrets or PII exposed?

Testability

  • How would you test this?
  • Are dependencies injectable?
  • Is there a test for the happy path? Edge cases?

Maintainability

  • Will the next developer understand this?
  • Is this doing too many things?
  • Is there duplication we could reduce?

Minimum Findings Enforcement

Reviews must meet a minimum weighted finding score of 3.0 (CRITICAL=3, HIGH=2, MEDIUM=1, LOW=0.5, INFORMATIONAL=0.25). If the initial review falls short, run the qe-devils-advocate agent as a meta-reviewer to find additional observations. Every review should have at least 3 actionable observations.


Agent-Assisted Reviews

// Comprehensive code review
await Task("Code Review", {
  prNumber: 123,
  checks: ['security', 'performance', 'testability', 'maintainability'],
  feedbackLevels: ['blocker', 'major', 'minor'],
  autoApprove: { maxBlockers: 0, maxMajor: 2 }
}, "qe-quality-analyzer");

// Security-focused review
await Task("Security Review", {
  prFiles: changedFiles,
  scanTypes: ['injection', 'auth', 'secrets', 'dependencies']
}, "qe-security-scanner");

// Test coverage review
await Task("Coverage Review", {
  prNumber: 123,
  requireNewTests: true,
  minCoverageDelta: 0
}, "qe-coverage-analyzer");

Agent Coordination Hints

Memory Namespace

aqe/code-review/
├── review-history/*     - Past review decisions
├── patterns/*           - Common issues by team/repo
├── feedback-templates/* - Reusable feedback
└── metrics/*            - Review turnaround time

Fleet Coordination

const reviewFleet = await FleetManager.coordinate({
  strategy: 'code-review',
  agents: [
    'qe-quality-analyzer',    // Logic, maintainability
    'qe-security-scanner',    // Security risks
    'qe-performance-tester',  // Performance issues
    'qe-coverage-analyzer'    // Test coverage
  ],
  topology: 'parallel'
});

Review Etiquette

✅ Do❌ Don't
"Have you considered...?""This is wrong"
Explain why it mattersJust say "fix this"
Acknowledge good codeOnly point out negatives
Suggest, don't demandBe condescending
Review < 400 linesReview 2000 lines at once

Related Skills


Remember

Prioritize feedback: 🔴 Blocker → 🟡 Major → 🟢 Minor → 💡 Suggestion. Focus on bugs and security, not style. Ask questions, don't command. Review < 400 lines at a time. Fast feedback (< 24h) beats thorough feedback.

With Agents: Agents automate security, performance, and coverage checks, freeing human reviewers to focus on logic and design. Use agents for consistent, fast initial review.

Skill Composition

  • Security concerns → Compose with /security-testing for security-focused review
  • Coverage check → Run /qe-coverage-analysis on changed files
  • Ship decision → Feed review results into /qe-quality-assessment

Gotchas

  • Agent reviews >400 lines at once and misses issues — chunk reviews to 200-400 lines maximum
  • Nitpicking style while missing logic bugs is the #1 agent review failure — prioritize correctness over formatting
  • Agent approves code that compiles but has subtle race conditions — always check shared state and async patterns
  • Review comments without suggested fixes are unhelpful — always include a proposed alternative
  • Agent doesn't check if the PR actually solves the linked issue — verify the stated problem is actually fixed

Repositorio GitHub

proffesor-for-testing/agentic-qe
Ruta: .claude/skills/code-review-quality
agenticqeagenticsfoundationagentsquality-engineering

Habilidades relacionadas

Verification & Quality Assurance

Otro

This skill provides automated quality verification for code and agent outputs using truth scoring and quality checks. It automatically rolls back changes that fall below a 0.95 accuracy threshold, ensuring codebase reliability. Use it for CI/CD integration and maintaining high-quality standards in development workflows.

Ver habilidad

Verification & Quality Assurance

Otro

This skill provides automated verification and quality assurance for code and agent outputs, including truth scoring and validation checks. It enables automatic rollback for failed quality checks and integrates with CI/CD pipelines. Use it to validate code changes before merging or to ensure the correctness of generated outputs.

Ver habilidad

testability-scoring

Otro

This skill provides AI-powered testability assessment for web applications using Playwright and optional Vibium integration. It evaluates applications against 10 intrinsic testability principles like Observability and Controllability to identify improvement areas. Use it when assessing software testability, evaluating test readiness, or generating testability reports.

Ver habilidad

refactoring-patterns

Otro

This Claude Skill applies safe refactoring patterns to improve code structure while preserving existing behavior. It's designed for cleaning up code, reducing technical debt, and enhancing maintainability through a disciplined test-first approach. The skill emphasizes small, incremental changes with continuous testing to ensure reliability during code transformations.

Ver habilidad