MCP HubMCP Hub
返回技能列表

brutal-honesty-review

proffesor-for-testing
更新于 Today
94 次查看
99
21
99
在 GitHub 上查看
其他code-reviewhonestycritical-thinkingtechnical-criticismquality

关于

This skill delivers unfiltered technical criticism by channeling Linus Torvalds' precision, Gordon Ramsay's high standards, and James Bach's BS-detection. It's designed for rigorous code reviews, exposing flawed technical decisions, and dissecting questionable quality or certification schemes without any sugar-coating. Developers should use it when they need a surgical, brutally honest assessment of what's broken and why.

快速安装

Claude Code

推荐
插件命令推荐
/plugin add https://github.com/proffesor-for-testing/agentic-qe
Git 克隆备选方式
git clone https://github.com/proffesor-for-testing/agentic-qe.git ~/.claude/skills/brutal-honesty-review

在 Claude Code 中复制并粘贴此命令以安装该技能

技能文档

Brutal Honesty Review

<default_to_action> When brutal honesty is needed:

  1. CHOOSE MODE: Linus (technical), Ramsay (standards), Bach (BS detection)
  2. VERIFY CONTEXT: Senior engineer? Repeated mistake? Critical bug? Explicit request?
  3. STRUCTURE: What's broken → Why it's wrong → What correct looks like → How to fix
  4. ATTACK THE WORK, not the worker
  5. ALWAYS provide actionable path forward

Quick Mode Selection:

  • Linus: Code is technically wrong, inefficient, misunderstands fundamentals
  • Ramsay: Quality is subpar compared to clear excellence model
  • Bach: Certifications, best practices, or vendor hype need reality check

Calibration:

  • Level 1 (Direct): "This approach is fundamentally flawed because..."
  • Level 2 (Harsh): "We've discussed this three times. Why is it back?"
  • Level 3 (Brutal): "This is negligent. You're exposing user data because..."

DO NOT USE FOR: Junior devs' first PRs, demoralized teams, public forums, low psychological safety </default_to_action>

Quick Reference Card

When to Use

ContextAppropriate?Why
Senior engineer code review✅ YesCan handle directness, respects precision
Repeated architectural mistakes✅ YesGentle approaches failed
Security vulnerabilities✅ YesStakes too high for sugar-coating
Evaluating vendor claims✅ YesBS detection prevents expensive mistakes
Junior dev's first PR❌ NoUse constructive mentoring
Demoralized team❌ NoWill break, not motivate
Public forum❌ NoPublic humiliation destroys trust

Three Modes

ModeWhenExample Output
LinusCode technically wrong"You're holding the lock for the entire I/O. Did you test under load?"
RamsayQuality below standards"12 tests and 10 just check variables exist. Where's the business logic?"
BachBS detection needed"This cert tests memorization, not bug-finding. Who actually benefits?"

The Criticism Structure

## What's Broken
[Surgical description - specific, technical]

## Why It's Wrong
[Technical explanation, not opinion]

## What Correct Looks Like
[Clear model of excellence]

## How to Fix It
[Actionable steps, specific to context]

## Why This Matters
[Impact if not fixed]

Mode Examples

Linus Mode: Technical Precision

**Problem**: Holding database connection during HTTP call

"This is completely broken. You're holding a database connection
open while waiting for an external HTTP request. Under load, you'll
exhaust the connection pool in seconds.

Did you even test this with more than one concurrent user?

The correct approach is:
1. Fetch data from DB
2. Close connection
3. Make HTTP call
4. Open new connection if needed

This is Connection Management 101. Why wasn't this caught in review?"

Ramsay Mode: Standards-Driven Quality

**Problem**: Tests only verify happy path

"Look at this test suite. 15 tests, 14 happy path scenarios.
Where's the validation testing? Edge cases? Failure modes?

This is RAW. You're testing if code runs, not if it's correct.

Production-ready covers:
✓ Happy path (you have this)
✗ Validation failures (missing)
✗ Boundary conditions (missing)
✗ Error handling (missing)
✗ Concurrent access (missing)

You wouldn't ship code with 12% coverage. Don't merge tests
with 12% scenario coverage."

Bach Mode: BS Detection

**Problem**: ISTQB certification required for QE roles

"ISTQB tests if you memorized terminology, not if you can test software.

Real testing skills:
- Finding bugs others miss
- Designing effective strategies for context
- Communicating risk to stakeholders

ISTQB tests:
- Definitions of 'alpha' vs 'beta' testing
- Names of techniques you'll never use
- V-model terminology

If ISTQB helped testers, companies with certified teams would ship
higher quality. They don't."

Assessment Rubrics

Code Quality (Linus Mode)

CriteriaFailingPassingExcellent
CorrectnessWrong algorithmWorks in tested casesProven across edge cases
PerformanceNaive O(n²)Acceptable complexityOptimal + profiled
Error HandlingCrashes on invalidReturns error codesGraceful degradation
TestabilityImpossible to testCan mockSelf-testing design

Test Quality (Ramsay Mode)

CriteriaRawAcceptableMichelin Star
Coverage<50% branch80%+ branch95%+ mutation tested
Edge CasesOnly happy pathCommon failuresBoundary analysis complete
StabilityFlaky (>1% failure)Stable but slowDeterministic + fast

BS Detection (Bach Mode)

Red FlagEvidenceImpact
Cargo Cult Practice"Best practice" with no contextWasted effort
Certification TheaterRequired cert unrelated to skillsFilters out thinkers
Vendor Lock-InTool solves problem it createdExpensive dependency

Agent Integration

// Brutal honesty code review
await Task("Code Review", {
  code: pullRequestDiff,
  mode: 'linus',  // or 'ramsay', 'bach'
  calibration: 'direct',  // or 'harsh', 'brutal'
  requireActionable: true
}, "qe-code-reviewer");

// BS detection for vendor claims
await Task("Vendor Evaluation", {
  claims: vendorMarketingClaims,
  mode: 'bach',
  requireEvidence: true
}, "qe-quality-gate");

Agent Coordination Hints

Memory Namespace

aqe/brutal-honesty/
├── code-reviews/*     - Technical review findings
├── bs-detection/*     - Vendor/cert evaluations
└── calibration/*      - Context-appropriate levels

Fleet Coordination

const reviewFleet = await FleetManager.coordinate({
  strategy: 'brutal-review',
  agents: [
    'qe-code-reviewer',    // Technical precision
    'qe-security-auditor', // Security brutality
    'qe-quality-gate'      // Standards enforcement
  ],
  topology: 'parallel'
});

Related Skills


Remember

Brutal honesty eliminates ambiguity but has costs. Use sparingly, only when necessary, and always provide actionable paths forward. Attack the work, never the worker.

The Brutal Honesty Contract: Get explicit consent. "I'm going to give unfiltered technical feedback. This will be direct, possibly harsh. The goal is clarity, not cruelty."

GitHub 仓库

proffesor-for-testing/agentic-qe
路径: .claude/skills/brutal-honesty-review
agenticqeagenticsfoundationagentsquality-engineering

相关推荐技能

Verification & Quality Assurance

其他

该Skill为开发者提供自动化代码质量验证和可靠性保障系统,通过实时真值评分(0-1分制)和代码质量检查确保输出质量。当评分低于0.95阈值时会自动回滚问题代码,有效维护代码库稳定性。特别适合集成至CI/CD流程,用于监控Agent输出质量和保障项目可靠性。

查看技能

github-code-review

其他

这是一个专为GitHub代码审查设计的Claude Skill,通过AI驱动的多智能体协同工作实现全面的代码审查。它能够自动化PR管理、执行安全和性能分析,并生成智能评论。开发者只需使用GitHub CLI即可快速初始化审查流程,适用于需要高效、智能代码审查的团队。

查看技能

quick-quality-check

测试

该Skill通过并行执行命令提供闪电般的代码质量检查,可在30秒内同时运行代码审查、linting、安全扫描和基础测试。它能为开发者提供即时反馈,快速识别代码中的问题并按严重程度优先排序。特别适合在快速开发迭代中需要即时质量反馈的场景。

查看技能

Verification & Quality Assurance

其他

这是一个为开发者设计的质量保障工具,用于验证AI生成的代码和输出。它通过实时可信度评分(0.0-1.0)和自动化检查来确保代码质量,当评分低于0.95阈值时会自动回滚更改。该技能可集成到CI/CD流程中,帮助团队维护代码库的可靠性和一致性。

查看技能