managing-database-partitions
Process use when you need to work with database partitioning. This skill provides table partitioning strategies with comprehensive guidance and automation. Trigger with phrases like "partition tables", "implement partitioning", or "optimize large tables".
Allowed Tools
Provided by Plugin
database-partition-manager
Database plugin for database-partition-manager
Installation
This skill is included in the database-partition-manager plugin:
/plugin install database-partition-manager@claude-code-plugins-plus
Click to copy
Instructions
Database Partition Manager
Overview
Implement and manage table partitioning for PostgreSQL and MySQL to improve query performance and simplify data lifecycle management on large tables. This skill covers range partitioning (by date or ID), list partitioning (by category or region), hash partitioning (for even distribution), and composite partitioning.
Prerequisites
- PostgreSQL 10+ (declarative partitioning) or MySQL 5.7+ (native partitioning)
- Database admin credentials with CREATE TABLE and ALTER TABLE permissions
psqlormysqlCLI for executing partition DDL- Table size metrics:
SELECT pgsizepretty(pgtotalrelationsize('tablename'))orSELECT datalength FROM informationschema.TABLES - Query patterns on the target table (especially WHERE clause columns used for filtering)
- Maintenance window availability for initial partition migration on existing tables
Instructions
- Identify partitioning candidates by finding tables that exceed 10GB or 100M rows, have time-based query patterns, or require periodic data purging. Query
pgstatuser_tablesto find tables with high sequential scan counts on large row sets.
- Select the partition key based on the most common query filter column. For time-series data, use the timestamp column. For multi-tenant data, use tenant_id. The partition key must appear in most WHERE clauses to enable partition pruning.
- Choose the partitioning strategy:
- Range: Best for time-series data. Create monthly or daily partitions. Queries filtering by date range scan only relevant partitions.
- List: Best for categorical data. Create one partition per category, region, or status value.
- Hash: Best for even distribution when no natural range exists. Distribute rows across N partitions using hash of the partition key.
- Composite: Combine range + list for multi-dimensional partitioning (e.g., range by date, then list by region).
- For PostgreSQL, create the partitioned parent table:
CREATE TABLE orders (id bigint, createdat timestamptz, ...) PARTITION BY RANGE (createdat). Then create child partitions:CREATE TABLE orders202401 PARTITION OF orders FOR VALUES FROM ('2024-01-01') TO ('2024-02-01').
- For MySQL, define partitions inline:
ALTER TABLE orders PARTITION BY RANGE (YEAR(createdat) * 100 + MONTH(createdat)) (PARTITION p202401 VALUES LESS THAN (202402), ...).
- Migrate data from an existing unpartitioned table to a partitioned table:
- Create the new partitioned table with identical schema
- Copy data in batches:
INSERT INTO orderspartitioned SELECT * FROM ordersold WHERE created_at BETWEEN ... AND ... - Verify row counts match between old and new tables
- Rename tables atomically:
ALTER TABLE orders RENAME TO ordersold; ALTER TABLE orderspartitioned RENAME TO orders;
- Create indexes on each partition. In PostgreSQL, indexes on the parent table automatically propagate to child partitions. Create the primary key and any secondary indexes on the partitioned table.
- Automate future partition creation with a scheduled script or cron job. For monthly range partitions, create the next 3 months of partitions in advance to prevent INSERT failures when a new month begins.
- Implement partition maintenance: drop or detach old partitions for data retention (
ALTER TABLE orders DETACH PARTITION orders202201), then archive or delete the detached partition. This is vastly faster thanDELETE FROM orders WHERE created_at < '2023-01-01'.
- Verify partition pruning works by running
EXPLAINon typical queries and confirming only relevant partitions are scanned. Look for "Partitions: 1/24" in the plan output indicating effective pruning.
Output
- Partition DDL scripts for creating partitioned tables and child partitions
- Data migration scripts for moving data from unpartitioned to partitioned tables
- Partition maintenance scripts for automated creation, detachment, and archival
- Partition pruning verification queries confirming optimizer uses partition elimination
- Cron job configurations for scheduled partition creation and cleanup
Error Handling
| Error | Cause | Solution |
|---|---|---|
no partition of relation "table" found for row |
INSERT targets a range with no matching partition | Create the missing partition; implement automated partition pre-creation for future ranges |
| Partition pruning not occurring | Query filter does not use the partition key, or uses a function on the key column | Rewrite query to filter directly on the partition key column; avoid wrapping partition key in functions |
| Slow data migration from unpartitioned table | Single large INSERT/SELECT locks the table and fills WAL | Migrate in batches by partition range; use pgrepack for online migration; increase maintenanceworkmem and maxwal_size |
| Foreign key references prevent partitioning | PostgreSQL does not support foreign keys referencing partitioned tables (pre-v12) | Upgrade to PostgreSQL 12+; or remove FK constraints and enforce referential integrity at application level |
| Too many partitions causing planner slowdown | Hundreds or thousands of child partitions degrade query planning time | Use wider partition ranges (monthly instead of daily); enable enablepartitionpruning; consider sub-partitioning instead of flat partitioning |
Examples
Monthly range partitioning for an events table: A 500GB events table with 2B rows partitioned by created_at into monthly partitions. Queries filtering by date range (last 7 days, last month) now scan only 1-2 partitions instead of the full table. Partition drop replaces a 4-hour DELETE operation with a sub-second DDL command for monthly data purges.
Hash partitioning for a sessions table: A sessions table with random UUID primary keys and no natural range column. Hash partition by session_id across 16 partitions to distribute I/O evenly. Parallel sequential scans across partitions improve full-table analytic queries by 8x on an 8-core server.
Composite partitioning for multi-region SaaS: Orders table partitioned first by range on created_at (monthly), then by list on region (us-east, us-west, eu, asia). Queries for "all US orders this month" prune to just 2 of 48 total partitions, reducing scan volume by 96%.
Resources
- PostgreSQL table partitioning: https://www.postgresql.org/docs/current/ddl-partitioning.html
- MySQL partitioning: https://dev.mysql.com/doc/refman/8.0/en/partitioning.html
- pgpartman extension (automated partition management): https://github.com/pgpartman/pgpartman
- Partition pruning internals: https://www.postgresql.org/docs/current/ddl-partitioning.html#DDL-PARTITION-PRUNING