api-response-optimization
About
This Claude Skill optimizes API performance by implementing caching strategies, response compression, and efficient payload structures. Use it when dealing with slow API responses, large payloads, or scaling bottlenecks to reduce network traffic and improve backend efficiency. It provides actionable guidance for eliminating unnecessary data and optimizing query performance.
Documentation
API Response Optimization
Overview
Fast API responses improve overall application performance and user experience. Optimization focuses on payload size, caching, and query efficiency.
When to Use
- Slow API response times
- High server CPU/memory usage
- Large response payloads
- Performance degradation
- Scaling bottlenecks
Instructions
1. Response Payload Optimization
// Inefficient response (unnecessary data)
GET /api/users/123
{
"id": 123,
"name": "John",
"email": "[email protected]",
"password_hash": "...", // ❌ Should never send
"ssn": "123-45-6789", // ❌ Sensitive data
"internal_id": "xyz",
"created_at": "2024-01-01T00:00:00Z",
"updated_at": "2024-01-02T00:00:00Z",
"meta_data": {...}, // ❌ Unused fields
"address": {
"street": "123 Main",
"city": "City",
"state": "ST",
"zip": "12345",
"geo": {...} // ❌ Not needed
}
}
// Optimized response (only needed fields)
GET /api/users/123
{
"id": 123,
"name": "John",
"email": "[email protected]"
}
// Results: 2KB → 100 bytes (20x smaller)
// Sparse fieldsets pattern
GET /api/users/123?fields=name,email
{
"id": 123,
"name": "John",
"email": "[email protected]"
}
2. Caching Strategies
HTTP Caching Headers:
Cache-Control:
Immutable assets: Cache-Control: public, max-age=31536000
API responses: Cache-Control: private, max-age=300
No cache: Cache-Control: no-store
Revalidate: Cache-Control: max-age=0, must-revalidate
ETag:
- Unique identifier for response version
- If-None-Match: return 304 if unchanged
- Saves bandwidth on unchanged data
Last-Modified:
- If-Modified-Since: return 304 if unchanged
- Simple versioning mechanism
---
Application-Level Caching:
Database Query Caching:
- Cache expensive queries
- TTL: 5-30 minutes
- Invalidate on write
- Tools: Redis, Memcached
Response Caching:
- Cache entire API responses
- Use Cache-Control headers
- Key: URL + query params
- TTL: Based on data freshness
Fragment Caching:
- Cache parts of response
- Combine multiple fragments
- Different TTL per fragment
---
Cache Invalidation:
Time-based (TTL):
- Simple: expires after time
- Risk: stale data
- Best for: Non-critical data
Event-based:
- Invalidate on write
- Immediate freshness
- Requires coordination
Hybrid:
- TTL + event invalidation
- Short TTL + invalidate on change
- Good balance
---
Implementation Example:
GET /api/users/123/orders
Authorization: Bearer token
Cache-Control: public, max-age=300
Response:
HTTP/1.1 200 OK
Cache-Control: public, max-age=300
ETag: "123abc"
Last-Modified: 2024-01-01
{data: [...]}
-- Next request within 5 minutes from cache
-- After 5 minutes, revalidate with ETag
-- If unchanged: 304 Not Modified
3. Compression & Performance
Compression:
gzip:
Ratio: 60-80% reduction
Format: text/html, application/json
Overhead: CPU (minor)
brotli:
Ratio: 20% better than gzip
Support: Modern browsers (95%)
Overhead: Higher CPU
Implementation:
- Enable in server
- Set Accept-Encoding headers
- Measure: Before/after sizes
- Monitor: CPU impact
---
Performance Optimization:
Pagination:
- Limit: 20-100 items per request
- Offset pagination: Simple, slow for large offsets
- Cursor pagination: Efficient, stable
- Implementation: Always use limit
Filtering:
- Server-side filtering
- Reduce response size
- Example: ?status=active
Sorting:
- Server-side only
- Index frequently sorted fields
- Limit sort keys to 1-2 fields
Eager Loading:
- Fetch related data in one query
- Avoid N+1 problem
- Example: /users?include=posts
---
Metrics & Monitoring:
Track:
- API response time (target: <200ms)
- Payload size (target: <100KB)
- Cache hit rate (target: >80%)
- Server CPU/memory
Tools:
- New Relic APM
- DataDog
- Prometheus
- Custom logging
Setup alerts:
- Response time >500ms
- Payload >500KB
- Cache miss spike
- Error rates
4. Optimization Checklist
Payload:
[ ] Remove sensitive data
[ ] Remove unused fields
[ ] Implement sparse fieldsets
[ ] Compress payload
[ ] Use appropriate status codes
Caching:
[ ] HTTP caching headers set
[ ] ETags implemented
[ ] Application cache configured
[ ] Cache invalidation strategy
[ ] Cache monitoring
Query Efficiency:
[ ] Database queries optimized
[ ] N+1 queries fixed
[ ] Joins optimized
[ ] Indexes in place
Compression:
[ ] gzip enabled
[ ] brotli enabled (modern)
[ ] Accept-Encoding headers
[ ] Content-Encoding responses
Monitoring:
[ ] Response time tracked
[ ] Payload size tracked
[ ] Cache metrics
[ ] Error rates
[ ] Alerts configured
Expected Improvements:
- Response time: 500ms → 100ms
- Payload size: 500KB → 50KB
- Server load: 80% CPU → 30%
- Concurrent users: 100 → 1000
Key Points
- Remove unnecessary data from responses
- Implement HTTP caching headers
- Use ETag for revalidation
- Paginate large result sets
- Enable gzip/brotli compression
- Monitor response times
- Cache expensive queries
- Implement sparse fieldsets
- Measure before and after
- Set up continuous monitoring
Quick Install
/plugin add https://github.com/aj-geddes/useful-ai-prompts/tree/main/api-response-optimizationCopy and paste this command in Claude Code to install this skill
GitHub 仓库
Related Skills
evaluating-llms-harness
TestingThis Claude Skill runs the lm-evaluation-harness to benchmark LLMs across 60+ standardized academic tasks like MMLU and GSM8K. It's designed for developers to compare model quality, track training progress, or report academic results. The tool supports various backends including HuggingFace and vLLM models.
langchain
MetaLangChain is a framework for building LLM applications using agents, chains, and RAG pipelines. It supports multiple LLM providers, offers 500+ integrations, and includes features like tool calling and memory management. Use it for rapid prototyping and deploying production systems like chatbots, autonomous agents, and question-answering services.
huggingface-accelerate
DevelopmentHuggingFace Accelerate provides the simplest API for adding distributed training to PyTorch scripts with just 4 lines of code. It offers a unified interface for multiple distributed training frameworks like DeepSpeed, FSDP, and DDP while handling automatic device placement and mixed precision. This makes it ideal for developers who want to quickly scale their PyTorch training across multiple GPUs or nodes without complex configuration.
nestjs
MetaThis skill provides NestJS development standards and architectural patterns for building domain-centric applications. It covers modular design, dependency injection, decorator patterns, and key framework features like controllers, services, middleware, and interceptors. Use it when developing NestJS applications, implementing APIs, configuring microservices, or integrating with databases.
