MCP HubMCP Hub
返回技能列表

prp-generator

matteocervelli
更新于 Today
8 次查看
10
10
在 GitHub 上查看
worddesign

关于

The prp-generator skill auto-activates to create structured Product Requirements Prompt (PRP) documents from synthesized design outputs. It provides comprehensive templates containing implementation plans, testing strategies, dependencies, and success criteria. Developers use it to transform architectural designs into actionable guidance for the implementation phase.

技能文档

Purpose

The prp-generator skill provides structured templates and methods for creating comprehensive Product Requirements Prompt (PRP) documents. PRPs serve as the primary input for Phase 3 (Implementation), containing all design information, library documentation, dependencies, implementation plans, testing strategies, and success criteria needed to implement features.

When to Use

This skill auto-activates when you:

  • Generate PRP documents from synthesized design
  • Create implementation guidance from architectural design
  • Structure design information for implementation phase
  • Define implementation plans with step-by-step guidance
  • Specify testing strategies and success criteria
  • Document architecture, libraries, and dependencies for implementers

Provided Capabilities

1. Structured PRP Generation

  • Standard Template: Consistent PRP structure across all features
  • Section Templates: Pre-defined templates for each PRP section
  • Completeness Checking: Ensure all required sections present
  • Format Validation: Validate markdown structure and linking

2. Implementation Planning

  • Phased Approach: Break implementation into logical phases
  • Step-by-Step Guidance: Detailed steps within each phase
  • Dependency Ordering: Ensure dependencies installed before use
  • Validation Checkpoints: Define validation at each phase

3. Testing Strategy Documentation

  • Test Types: Unit, integration, end-to-end testing approaches
  • Test Coverage: Specify coverage targets per component
  • Mock Strategies: Document how to mock library dependencies
  • Test Data: Provide test data and fixture guidance

4. Success Criteria Definition

  • Acceptance Criteria: Map from requirements analysis
  • Implementation Criteria: Code-level success criteria
  • Testing Criteria: Test coverage and passing requirements
  • Documentation Criteria: Required documentation artifacts

5. Architecture Documentation

  • Component Diagrams: Visual architecture representations
  • Data Models: Pydantic schemas with validations
  • API Contracts: Endpoint specifications with examples
  • Data Flow: How data moves through system

Usage Guide

Step 1: Load Inputs

Load all required inputs for PRP generation:

Required Inputs:

  1. Analysis Document: /docs/implementation/analysis/feature-{issue-number}-analysis.md

    • Requirements (functional and non-functional)
    • Security considerations
    • Technical stack requirements
    • Dependencies
    • Scope definition
    • Identified risks
  2. Synthesized Design: Output from design-synthesizer skill

    • Component architecture with library mappings
    • Data models with library validation patterns
    • API contracts with library integration
    • Library integration strategies
    • Dependency resolution
    • Implementation path
    • Design decisions
    • Identified issues/gaps

Loading Pattern:

# Read analysis document
analysis = read_document("/docs/implementation/analysis/feature-{issue-number}-analysis.md")

# Parse synthesized design (from synthesis step)
design = parse_synthesized_design()

# Extract key information
requirements = extract_requirements(analysis)
architecture = extract_architecture(design)
libraries = extract_libraries(design)
dependencies = extract_dependencies(design)
implementation_plan = extract_implementation_plan(design)

Step 2: Create PRP Header

Use prp-template.md header section:

# Product Requirements Prompt: [Feature Name] (Issue #{issue-number})

**Date**: [YYYY-MM-DD]
**Designer**: Design Orchestrator (Claude Code)
**Issue**: #{issue-number} - [Issue Title]
**Analysis Document**: /docs/implementation/analysis/feature-{issue-number}-analysis.md

---

**Status**: Ready for Implementation (Phase 3)
**Complexity**: [Low / Medium / High]
**Estimated Effort**: [Hours/Days]
**Priority**: [P0 / P1 / P2 / P3]

---

Complexity Guidelines:

  • Low: Single component, <3 files, well-documented libraries
  • Medium: Multiple components, 3-10 files, some custom logic
  • High: Complex architecture, >10 files, custom integrations

Step 3: Write Executive Summary

Provide 2-3 paragraph overview of the design:

Executive Summary Template:

## Executive Summary

This document provides comprehensive implementation guidance for [feature name],
which [high-level description of what feature does and why it's valuable].

**Architectural Approach**: [Brief description of architecture - e.g., "Service layer
pattern with FastAPI endpoints, SQLAlchemy ORM, and Redis caching"]

**Key Libraries**: [List 3-5 primary libraries - e.g., "FastAPI 0.95.0 for API
framework, SQLAlchemy 2.0.0 for ORM, PassLib 1.7.4 for password hashing"]

**Implementation Strategy**: [Brief description of implementation phases - e.g.,
"4-phase approach: (1) Data models and repository, (2) Service layer business logic,
(3) API endpoints, (4) Testing and validation"]

**Key Considerations**: [1-2 important considerations - e.g., "Security-first
approach with OWASP compliance, performance target of <200ms API response time"]

Example:

## Executive Summary

This document provides comprehensive implementation guidance for user authentication
system, which enables secure user login, session management, and access control for
the application.

**Architectural Approach**: Service layer pattern with FastAPI RESTful API endpoints,
SQLAlchemy ORM for user persistence, PassLib for secure password hashing, and JWT
tokens for session management.

**Key Libraries**: FastAPI 0.95.0 (API framework), SQLAlchemy 2.0.0 (ORM), PassLib
1.7.4 (password hashing), python-jose 3.3.0 (JWT tokens), pydantic 1.10.7 (validation).

**Implementation Strategy**: 4-phase approach: (1) User data model and repository
with SQLAlchemy, (2) Authentication service with PassLib and JWT, (3) FastAPI
endpoints with pydantic validation, (4) Comprehensive testing with pytest and mocks.

**Key Considerations**: OWASP Top 10 compliance for authentication vulnerabilities,
bcrypt with cost factor 12 for password hashing, rate limiting for brute force
protection, secure session token management with expiration.

Step 4: Document Requirements Reference

Link to analysis document and summarize key requirements:

Requirements Reference Template:

## Requirements Reference

**Source**: [Link to analysis document]

### Functional Requirements (Summary)
- **[FR-001]**: [Brief description]
- **[FR-002]**: [Brief description]
...

### Non-Functional Requirements (Summary)
- **Performance**: [Key performance requirements]
- **Security**: [Key security requirements]
- **Usability**: [Key usability requirements]
- **Scalability**: [Key scalability requirements]

### Acceptance Criteria (Key Points)
- [ ] [Top acceptance criterion 1]
- [ ] [Top acceptance criterion 2]
- [ ] [Top acceptance criterion 3]

**Full Requirements**: See analysis document for complete requirements, security
assessment, and risk analysis.

Step 5: Document Architecture Design

Use synthesized architecture design:

Architecture Design Template:

## Architecture Design

### Component Overview

[ASCII diagram or description of component architecture]

### Components

#### Component: [ComponentName]
**Purpose**: [What this component does]
**Responsibilities**:
- [Responsibility 1]
- [Responsibility 2]

**Library Integration**:
- **Primary Library**: [LibraryName] v[Version]
- **APIs Used**: [Specific library APIs/classes]
- **Pattern**: [Integration pattern from design-patterns.md]
- **Dependencies**: [Required dependencies]

**Implementation Notes**:
- [Important implementation detail 1]
- [Important implementation detail 2]

**Code Example**:
```python
# Example from library documentation (libraries-{issue-number}.md, Example X.Y)
[Code snippet showing library usage pattern]

[Repeat for each component]

Data Models

Model: [ModelName]

Purpose: [What this model represents] Fields:

from pydantic import BaseModel, Field
from [library] import [ValidationMixin]

class [ModelName](BaseModel):
    field1: type = Field(..., description="...")  # Validation
    field2: type = Field(default=..., ge=..., le=...)  # Range validation
    # ... more fields

Validation Rules:

  • [Field1]: [Validation description]
  • [Field2]: [Validation description]

Relationships:

  • [Related Model]: [Relationship type and description]

Library Integration:

  • Validation: [Library providing validation - e.g., Pydantic validators]
  • Serialization: [How model is serialized - e.g., Pydantic .dict()]
  • Storage: [ORM pattern - e.g., SQLAlchemy declarative base]

[Repeat for each model]

API Contracts

Endpoint: [METHOD] [/path]

Purpose: [What this endpoint does]

Request:

class [RequestModel](BaseModel):
    field1: type
    field2: type

Response:

class [ResponseModel](BaseModel):
    field1: type
    field2: type

Status Codes:

  • 200: Success - [Description]
  • 400: Bad Request - [Description]
  • 401: Unauthorized - [Description]
  • 404: Not Found - [Description]
  • 500: Internal Server Error - [Description]

Implementation:

# FastAPI endpoint pattern from library docs
@app.[method]("/path", response_model=[ResponseModel])
def endpoint([params: RequestModel]):
    # Implementation using service layer
    ...

[Repeat for each endpoint]

Data Flow

[Describe how data flows through system]

Example Flow: User Authentication

1. Client sends POST /auth/login with email/password
2. FastAPI endpoint validates request (Pydantic)
3. AuthenticationService.authenticate() called
4. UserRepository.get_by_email() queries database (SQLAlchemy)
5. PassLib.verify() checks password hash
6. JWT token generated (python-jose)
7. Token returned to client

Error Handling Strategy

Library Exception Mapping:

Library ExceptionArchitectural ExceptionHTTP StatusUser Message
[LibraryException][AppException][Code][Message]

Error Handling Pattern:

try:
    # Library operation
except LibraryException as e:
    # Map to architectural exception
    raise AppException(...) from e

### Step 6: Document Library Documentation

Use library information from Documentation Researcher:

**Library Documentation Template**:
```markdown
## Library Documentation

### Library: [LibraryName] v[Version]

**Purpose**: [What library does]
**Documentation**: [Link to official docs]
**Repository**: [Link to GitHub/source]

**Installation**:
```bash
pip install [library-name]==[version]

Key APIs Used:

  • [API1]: [Description and purpose]
  • [API2]: [Description and purpose]

Integration Pattern:

# Example from library documentation
[Code example showing how to use library in this project]

Configuration:

# Configuration required for this library
[Configuration code or settings]

Best Practices (from library docs):

  • [Best practice 1]
  • [Best practice 2]

Known Issues:

  • [Issue 1 and workaround]
  • [Issue 2 and workaround]

Testing Notes:

  • Mocking: [How to mock this library in tests]
  • Test Containers: [If integration tests need real library instance]

[Repeat for each library]

Library Comparison (if alternatives considered)

LibraryVersionProsConsDecision
[Lib A][Ver][Pros][Cons]✅ Chosen
[Lib B][Ver][Pros][Cons]❌ Not chosen

Selection Rationale: [Why chosen library was selected]


### Step 7: Document Dependencies

Use dependency information from Dependency Manager:

**Dependencies Template**:
```markdown
## Dependencies

### Dependency Tree

[Full dependency tree with versions] [library-a]==[version] ├── [dep-1]==[version] │ └── [sub-dep]==[version] ├── [dep-2]==[version] └── [optional-dep] (optional)


### Installation Commands

**Install all dependencies**:
```bash
# Install using pip
pip install -r requirements.txt

# Or using uv (faster)
uv pip install -r requirements.txt

Install in order (if order matters):

# Step 1: Install base dependencies
pip install [base-deps]

# Step 2: Install main libraries
pip install [main-libs]

# Step 3: Install optional dependencies
pip install [optional-deps]

Dependency Details

Dependency: [dep-name] v[version]

Purpose: [What this dependency provides] Required By: [Libraries that need this dependency] Version Constraint: [Why this specific version] Alternatives: [Alternative versions/libraries considered]

[Repeat for key dependencies]

Compatibility Notes

Python Version: [Required Python version] Operating System: [OS compatibility notes] Database: [Database version requirements] Other System Dependencies: [System-level requirements]

Dependency Conflicts and Resolutions

[If any conflicts were found and resolved]

Conflict: [Description]

  • Library A needs: [Constraint A]
  • Library B needs: [Constraint B]
  • Resolution: [How conflict was resolved]
  • Pinned Version: [Final version chosen]

### Step 8: Create Implementation Plan

Use `prp-template.md` implementation plan section:

**Implementation Plan Template**:
```markdown
## Implementation Plan

### Overview

This implementation follows a [N]-phase approach, progressing from foundation
(data models) to core functionality to integration to testing. Each phase builds
on the previous phase and has validation checkpoints.

### Phase 1: Foundation ([Estimated Time])

**Goal**: Create data models, repositories, and core utilities

**Tasks**:
1. **Data Models** ([Time])
   - [ ] Create [Model1] Pydantic schema with validations
   - [ ] Create [Model2] Pydantic schema with validations
   - [ ] Add [Library]-specific validators
   - [ ] Write unit tests for model validation

2. **Repository Layer** ([Time])
   - [ ] Create [Repository1] with CRUD operations
   - [ ] Implement [Library] ORM mappings
   - [ ] Add query methods for common operations
   - [ ] Write repository unit tests with mocks

3. **Configuration** ([Time])
   - [ ] Create configuration classes (Pydantic BaseSettings)
   - [ ] Add environment variable loading
   - [ ] Configure [Library1], [Library2]
   - [ ] Write configuration validation tests

**Validation Checkpoint**:
- [ ] All data models validate correctly
- [ ] All repository operations work with test database
- [ ] Configuration loads from environment
- [ ] Phase 1 tests pass (pytest)

### Phase 2: Core Implementation ([Estimated Time])

**Goal**: Implement business logic services using libraries

**Tasks**:
1. **Service Layer** ([Time])
   - [ ] Create [Service1] with business logic
   - [ ] Integrate [Library1] for [functionality]
   - [ ] Implement error handling and exception mapping
   - [ ] Add logging and monitoring
   - [ ] Write service unit tests with mocked dependencies

2. **Library Integration** ([Time])
   - [ ] Set up [Library2] client/connection
   - [ ] Implement [Library2] operations
   - [ ] Add retry logic for transient failures
   - [ ] Write integration tests with test containers

**Validation Checkpoint**:
- [ ] All service methods work correctly
- [ ] Library integrations functioning
- [ ] Error handling covers all cases
- [ ] Phase 2 tests pass (pytest)

### Phase 3: Integration ([Estimated Time])

**Goal**: Connect components and create API endpoints

**Tasks**:
1. **API Endpoints** ([Time])
   - [ ] Create [Endpoint1] with [Method] [/path]
   - [ ] Add request/response validation (Pydantic)
   - [ ] Integrate with service layer
   - [ ] Add authentication/authorization
   - [ ] Write endpoint integration tests

2. **Component Integration** ([Time])
   - [ ] Connect [Component A] to [Component B]
   - [ ] Implement data flow [Flow description]
   - [ ] Add transaction handling
   - [ ] Write integration tests for full flows

3. **Error Handling** ([Time])
   - [ ] Add global exception handlers
   - [ ] Map library exceptions to HTTP status codes
   - [ ] Implement error response formatting
   - [ ] Test error scenarios

**Validation Checkpoint**:
- [ ] All API endpoints respond correctly
- [ ] End-to-end flows work
- [ ] Error handling covers all endpoints
- [ ] Phase 3 tests pass (pytest)

### Phase 4: Testing & Validation ([Estimated Time])

**Goal**: Comprehensive testing and final validation

**Tasks**:
1. **Unit Test Coverage** ([Time])
   - [ ] Achieve [X]% unit test coverage
   - [ ] All components have isolated unit tests
   - [ ] All edge cases covered
   - [ ] Mock all external dependencies

2. **Integration Testing** ([Time])
   - [ ] Test all API endpoints (end-to-end)
   - [ ] Test library integrations with real instances
   - [ ] Test error scenarios and recovery
   - [ ] Performance testing ([targets])

3. **Security Validation** ([Time])
   - [ ] OWASP Top 10 verification
   - [ ] Input validation testing
   - [ ] Authentication/authorization testing
   - [ ] Secrets management verification

4. **Documentation** ([Time])
   - [ ] Update API documentation
   - [ ] Add code comments for complex logic
   - [ ] Create user guide (if needed)
   - [ ] Update README with new feature

**Final Validation**:
- [ ] All acceptance criteria met
- [ ] All tests pass (pytest)
- [ ] Coverage targets achieved
- [ ] Security validation complete
- [ ] Documentation complete
- [ ] Code review passed

### Phase 5: Deployment ([Estimated Time])

**Goal**: Deploy to production environment

**Tasks**:
1. **Pre-Deployment** ([Time])
   - [ ] Review deployment checklist
   - [ ] Verify environment configuration
   - [ ] Database migrations tested
   - [ ] Rollback plan documented

2. **Deployment** ([Time])
   - [ ] Deploy to staging environment
   - [ ] Run smoke tests in staging
   - [ ] Deploy to production
   - [ ] Monitor for errors

3. **Post-Deployment** ([Time])
   - [ ] Verify production functionality
   - [ ] Monitor performance metrics
   - [ ] Check logs for errors
   - [ ] Update documentation

**Deployment Validation**:
- [ ] Feature working in production
- [ ] No critical errors in logs
- [ ] Performance meets targets
- [ ] Monitoring dashboards updated

Step 9: Define Testing Strategy

Use prp-template.md testing section:

Testing Strategy Template:

## Testing Strategy

### Overview

Comprehensive testing approach with [X]% code coverage target, including unit tests
(with mocks), integration tests (with test containers), and end-to-end tests (full
API flow).

### Unit Testing

**Scope**: Individual functions, classes, methods in isolation

**Framework**: pytest

**Coverage Target**: [X]% for all modules

**Mocking Strategy**:
- Mock [Library1] using unittest.mock
- Mock [Library2] using [library-specific test fixtures]
- Mock database using SQLAlchemy in-memory database

**Test Structure**:

tests/ ├── unit/ │ ├── test_models.py # Data model validation tests │ ├── test_repositories.py # Repository logic tests (mocked DB) │ ├── test_services.py # Service logic tests (mocked deps) │ └── test_validators.py # Custom validator tests


**Example Unit Test**:
```python
# Test with mocked library
from unittest.mock import Mock
import pytest

def test_service_method():
    # Mock library dependency
    mock_lib = Mock()
    mock_lib.method.return_value = "expected"

    # Test service with mock
    service = MyService(mock_lib)
    result = service.do_something()

    assert result == "processed_expected"
    mock_lib.method.assert_called_once()

Key Test Cases:

  • [Component1]: Test [key functionality]
  • [Component2]: Test [error handling]
  • [Model1]: Test [validation rules]

Integration Testing

Scope: Components working together with real library instances

Framework: pytest with testcontainers

Test Containers Used:

  • PostgreSQL: testcontainers.postgres.PostgresContainer
  • Redis: testcontainers.redis.RedisContainer

Test Structure:

tests/
├── integration/
│   ├── test_database_operations.py   # Real DB tests
│   ├── test_cache_operations.py      # Real Redis tests
│   └── test_library_integrations.py  # Real library usage

Example Integration Test:

from testcontainers.postgres import PostgresContainer

def test_repository_integration():
    with PostgresContainer("postgres:15") as postgres:
        # Create real database engine
        engine = create_engine(postgres.get_connection_url())
        # Test with real database
        ...

Key Test Cases:

  • [Integration1]: Test [data flow]
  • [Integration2]: Test [library interaction]

End-to-End Testing

Scope: Complete API flows from request to response

Framework: pytest with TestClient (FastAPI) or requests

Test Structure:

tests/
├── e2e/
│   ├── test_api_flows.py         # Complete API workflows
│   ├── test_authentication.py    # Auth flows
│   └── test_error_scenarios.py   # Error handling

Example E2E Test:

from fastapi.testclient import TestClient

def test_user_creation_flow():
    client = TestClient(app)

    # Full flow test
    response = client.post("/users", json={"email": "[email protected]"})
    assert response.status_code == 201

    user_id = response.json()["id"]
    get_response = client.get(f"/users/{user_id}")
    assert get_response.status_code == 200

Key Test Cases:

  • [Flow1]: Test [complete user journey]
  • [Flow2]: Test [error recovery]

Performance Testing

Scope: Verify performance meets non-functional requirements

Tools: pytest-benchmark, locust

Target Metrics:

  • API response time: [target] (95th percentile)
  • Database query time: [target]
  • Cache hit rate: [target]

Test Cases:

  • Benchmark [operation] performance
  • Load test [endpoint] with [N] concurrent requests

Security Testing

Scope: Verify security controls and OWASP compliance

Test Cases:

  • SQL injection prevention (parameterized queries)
  • XSS prevention (output sanitization)
  • CSRF protection (tokens)
  • Authentication bypass attempts
  • Authorization boundary testing
  • Secrets not in logs or responses

Test Data Management

Test Fixtures:

# Shared test fixtures
@pytest.fixture
def sample_user():
    return User(id=1, email="[email protected]", name="Test User")

@pytest.fixture
def mock_database():
    # Create in-memory database for testing
    ...

Test Data Location: tests/fixtures/

Running Tests

All Tests:

pytest

Unit Tests Only:

pytest tests/unit/

Integration Tests Only:

pytest tests/integration/

With Coverage:

pytest --cov=src --cov-report=html

Coverage Report: htmlcov/index.html

Continuous Integration

CI Pipeline (GitHub Actions / GitLab CI):

  1. Run linting (black, mypy)
  2. Run unit tests
  3. Run integration tests (with test containers)
  4. Generate coverage report
  5. Fail if coverage < [X]%

### Step 10: Document Success Criteria

Use acceptance criteria from analysis and implementation-specific criteria:

**Success Criteria Template**:
```markdown
## Success Criteria

### Acceptance Criteria (from Requirements)

From analysis document (feature-{issue-number}-analysis.md):

- [ ] **AC-001**: [Acceptance criterion from analysis]
- [ ] **AC-002**: [Acceptance criterion from analysis]
- [ ] **AC-003**: [Acceptance criterion from analysis]

### Implementation Criteria

- [ ] **IC-001**: All data models defined with Pydantic validation
- [ ] **IC-002**: All repository CRUD operations implemented with [ORM]
- [ ] **IC-003**: All service methods implemented with error handling
- [ ] **IC-004**: All API endpoints implemented with [Framework]
- [ ] **IC-005**: All library integrations working ([Library1], [Library2])

### Testing Criteria

- [ ] **TC-001**: Unit test coverage ≥ [X]%
- [ ] **TC-002**: All integration tests passing
- [ ] **TC-003**: All end-to-end API tests passing
- [ ] **TC-004**: Performance tests meet targets ([metric] < [target])
- [ ] **TC-005**: Security tests passing (OWASP compliance)

### Code Quality Criteria

- [ ] **QC-001**: Linting passes (black, mypy)
- [ ] **QC-002**: No critical security vulnerabilities (bandit/safety)
- [ ] **QC-003**: Code review approved
- [ ] **QC-004**: All functions have docstrings
- [ ] **QC-005**: Complex logic has explanatory comments

### Documentation Criteria

- [ ] **DC-001**: API endpoints documented (OpenAPI/Swagger)
- [ ] **DC-002**: README updated with new feature
- [ ] **DC-003**: User guide created (if user-facing)
- [ ] **DC-004**: Code comments for complex logic
- [ ] **DC-005**: CHANGELOG updated

### Deployment Criteria

- [ ] **DEP-001**: Feature deployed to staging
- [ ] **DEP-002**: Smoke tests passing in staging
- [ ] **DEP-003**: Feature deployed to production
- [ ] **DEP-004**: Production monitoring dashboards updated
- [ ] **DEP-005**: No critical errors in first 24 hours

### Definition of Done

All of the following must be true:
- ✅ All acceptance criteria met
- ✅ All implementation criteria met
- ✅ All testing criteria met
- ✅ All code quality criteria met
- ✅ All documentation criteria met
- ✅ All deployment criteria met
- ✅ Feature reviewed and approved
- ✅ Feature in production and stable

Step 11: Validate PRP Completeness

Use prp-template.md validation checklist:

PRP Completeness Checklist:

  • Header with metadata (date, designer, issue, analysis link)
  • Executive summary (2-3 paragraphs)
  • Requirements reference with key points
  • Architecture design (components, data models, APIs, data flow, error handling)
  • Library documentation (all libraries with examples and best practices)
  • Dependencies (tree, installation, compatibility)
  • Implementation plan (phased with tasks and validation checkpoints)
  • Testing strategy (unit, integration, e2e, performance, security)
  • Documentation requirements (what docs to create)
  • Success criteria (acceptance, implementation, testing, quality, docs, deployment)
  • Risks & mitigations (from synthesis)

Step 12: Write PRP to File

Save PRP to designated location:

File Path: /docs/implementation/prp/feature-{issue-number}-prp.md

Final PRP Structure:

# Product Requirements Prompt: [Feature Name] (Issue #{issue-number})
[Header with metadata]
## Executive Summary
[2-3 paragraphs]
## Requirements Reference
[Key requirements from analysis]
## Architecture Design
[Components, models, APIs, data flow, error handling]
## Library Documentation
[All libraries with examples]
## Dependencies
[Tree, installation, compatibility]
## Implementation Plan
[Phases with tasks and validation]
## Testing Strategy
[Unit, integration, e2e, performance, security]
## Documentation Requirements
[What to document]
## Success Criteria
[Acceptance, implementation, testing, quality, docs, deployment]
## Risks & Mitigations
[From synthesis and analysis]
---
**PRP Complete**: [Date/Time]
**Ready for Phase 3**: Implementation

Best Practices

1. Be Comprehensive But Concise

  • Include all necessary information
  • Avoid redundancy
  • Link to other documents for details
  • Use examples liberally

2. Make It Actionable

  • Every section should guide implementer
  • Provide code examples
  • Specify exact commands
  • Define clear validation checkpoints

3. Ensure Traceability

  • Link requirements to design elements
  • Link design to implementation tasks
  • Link tasks to test cases
  • Link tests to acceptance criteria

4. Validate Completeness

  • Use prp-template.md checklist
  • Verify all required sections present
  • Check all links work
  • Ensure all code examples valid

5. Consider the Implementer

  • Write for someone unfamiliar with design discussions
  • Provide context and rationale
  • Explain trade-offs and decisions
  • Anticipate questions

Resources

prp-template.md

Complete PRP template with:

  • Standard structure
  • Section templates
  • Markdown formatting
  • Validation checklist

prp-examples.md

Example PRPs showing:

  • Simple feature PRP
  • Complex feature PRP
  • Multi-component PRP
  • API-focused PRP
  • Data-focused PRP

Example Usage

See prp-examples.md for complete PRP examples.

Integration

This skill is used by:

  • design-orchestrator agent during Phase 2: Design & Planning
  • Activates automatically when generating PRP from synthesized design
  • Output feeds into Phase 3: Implementation (Implementation Specialist)

Version: 2.0.0 Auto-Activation: Yes (when generating PRP documents) Phase: 2 (Design & Planning) Created: 2025-10-29

快速安装

/plugin add https://github.com/matteocervelli/llms/tree/main/prp-generator

在 Claude Code 中复制并粘贴此命令以安装该技能

GitHub 仓库

matteocervelli/llms
路径: .claude/skills/prp-generator

相关推荐技能

langchain

LangChain是一个用于构建LLM应用程序的框架,支持智能体、链和RAG应用开发。它提供多模型提供商支持、500+工具集成、记忆管理和向量检索等核心功能。开发者可用它快速构建聊天机器人、问答系统和自主代理,适用于从原型验证到生产部署的全流程。

查看技能

go-test

go-test Skill为Go开发者提供全面的测试指导,涵盖单元测试、性能基准测试和集成测试的最佳实践。它能帮助您正确实现表驱动测试、子测试组织、mock接口和竞态检测,同时指导测试覆盖率分析和性能基准测试。当您编写_test.go文件、设计测试用例或优化测试策略时,这个Skill能确保您遵循Go语言的标准测试惯例。

查看技能

project-structure

这个Skill为开发者提供全面的项目目录结构设计指南和最佳实践。它涵盖了多种项目类型包括monorepo、前后端框架、库和扩展的标准组织结构。帮助团队创建可扩展、易维护的代码架构,特别适用于新项目设计、遗留项目迁移和团队规范制定。

查看技能

issue-documentation

该Skill为开发者提供标准化的issue文档模板和指南,适用于创建bug报告、GitHub/Linear/Jira问题等场景。它能系统化地记录问题状况、复现步骤、根本原因、解决方案和影响范围,确保团队沟通清晰高效。通过实施主流问题跟踪系统的最佳实践,帮助开发者生成结构完整的故障排除文档和事件报告。

查看技能