generating-database-seed-data
About
This skill generates realistic database seed scripts and test data using Faker libraries for development and testing. It maintains relational integrity and allows configurable data volumes when populating databases. Use it for quickly creating sample data with triggers like "seed database" or "generate test data".
Quick Install
Claude Code
Recommended/plugin add https://github.com/jeremylongshore/claude-code-plugins-plusgit clone https://github.com/jeremylongshore/claude-code-plugins-plus.git ~/.claude/skills/generating-database-seed-dataCopy and paste this command in Claude Code to install this skill
Documentation
Overview
This skill automates the creation of database seed scripts, populating your database with realistic and consistent test data. It leverages Faker libraries to generate diverse and believable data, ensuring relational integrity and configurable data volumes.
How It Works
- Analyze Schema: Claude analyzes the database schema to understand table structures and relationships.
- Generate Data: Using Faker libraries, Claude generates realistic data for each table, respecting data types and constraints.
- Maintain Relationships: Claude ensures foreign key relationships are maintained, creating consistent and valid data across tables.
- Create Seed Script: Claude generates a database seed script (e.g., SQL, JavaScript) containing the generated data.
When to Use This Skill
This skill activates when you need to:
- Populate a development database with realistic data.
- Create a seed script for automated database setup.
- Generate test data for application testing.
- Demonstrate an application with pre-populated data.
Examples
Example 1: Populating a User Database
User request: "Create a seed script to populate my users table with 50 realistic users."
The skill will:
- Analyze the 'users' table schema (name, email, password, etc.).
- Generate 50 sets of realistic user data using Faker libraries.
- Create a SQL seed script to insert the generated user data into the 'users' table.
Example 2: Seeding a Blog Database
User request: "Generate test data for my blog database, including posts, comments, and users."
The skill will:
- Analyze the 'posts', 'comments', and 'users' table schemas and their relationships.
- Generate realistic data for each table, ensuring foreign key relationships are maintained (e.g., comments linked to posts, posts linked to users).
- Create a seed script (e.g., JavaScript with TypeORM) to insert the generated data into the database.
Best Practices
- Data Volume: Start with a small data volume and gradually increase it to avoid performance issues.
- Data Consistency: Ensure the Faker libraries used are appropriate for the data types and formats required by your database.
- Idempotency: Design your seed scripts to be idempotent, so they can be run multiple times without causing errors or duplicate data.
Integration
This skill integrates well with database migration tools and frameworks, allowing you to automate the entire database setup process, including schema creation and data seeding. It can also be used in conjunction with testing frameworks to generate realistic test data for automated testing.
GitHub Repository
Related Skills
sglang
MetaSGLang is a high-performance LLM serving framework that specializes in fast, structured generation for JSON, regex, and agentic workflows using its RadixAttention prefix caching. It delivers significantly faster inference, especially for tasks with repeated prefixes, making it ideal for complex, structured outputs and multi-turn conversations. Choose SGLang over alternatives like vLLM when you need constrained decoding or are building applications with extensive prefix sharing.
evaluating-llms-harness
TestingThis Claude Skill runs the lm-evaluation-harness to benchmark LLMs across 60+ standardized academic tasks like MMLU and GSM8K. It's designed for developers to compare model quality, track training progress, or report academic results. The tool supports various backends including HuggingFace and vLLM models.
llamaguard
OtherLlamaGuard is Meta's 7-8B parameter model for moderating LLM inputs and outputs across six safety categories like violence and hate speech. It offers 94-95% accuracy and can be deployed using vLLM, Hugging Face, or Amazon SageMaker. Use this skill to easily integrate content filtering and safety guardrails into your AI applications.
langchain
MetaLangChain is a framework for building LLM applications using agents, chains, and RAG pipelines. It supports multiple LLM providers, offers 500+ integrations, and includes features like tool calling and memory management. Use it for rapid prototyping and deploying production systems like chatbots, autonomous agents, and question-answering services.
