Summary
Most developers use AI to build features. Advanced developers use AI to build infrastructure that builds features. Elite developers build infrastructure that builds infrastructure. This article explores the levels of automation and shows how to use Claude Code to create self-improving development systems that compound productivity exponentially.
The Problem
Developers use AI coding agents to complete individual tasks (writing functions, fixing bugs, adding features) instead of building automation infrastructure. This creates linear productivity: each task requires AI assistance. The opportunity cost is massive—time spent on repetitive tasks could be spent building tools that eliminate those tasks forever.
The Solution
Use Claude Code to build infrastructure that builds infrastructure (“building the factory”). Progress through automation levels: (0) manual coding → (1) AI-assisted features → (2) building tools with AI → (3) building tools that build tools. Focus on high-leverage infrastructure: custom MCP servers, development bootstrap tools, evaluation suites, and Claude Code orchestration systems. The result: exponential productivity where tools multiply your capacity automatically.
The Problem
You’re using Claude Code to build features. Each day, you ask it to:
- “Add authentication to the API”
- “Fix the broken database migration”
- “Implement the new dashboard component”
- “Debug the failing tests”
Claude Code completes these tasks efficiently. You’re productive. But you’re thinking linearly.
Linear vs Exponential Productivity
Linear productivity (most developers):
Day 1: Use AI to build Feature A (8 hours saved)
Day 2: Use AI to build Feature B (8 hours saved)
Day 3: Use AI to build Feature C (8 hours saved)
...
Total: 8N hours saved over N days
Exponential productivity (building the factory):
Week 1: Build tool that generates Feature scaffolding (16 hours invested)
→ Saves 2 hours per feature going forward
Week 2: Build 3 more tools (24 hours invested)
→ Saves 6 hours per day
Week 4: Build 10 tools (40 hours invested)
→ Most work happens automatically
Month 2: Tools build tools
→ Exponential growth
Month 6: Self-improving infrastructure
→ Work happens while you sleep
The key difference: Linear productivity trades time for output. Exponential productivity invests time in capacity expansion.
The Opportunity Cost
Every hour you spend asking Claude Code to “add feature X” is an hour you could spend asking it to “build a tool that generates features like X automatically.”
Scenario A: Use AI for 100 features
Time: 100 × 2 hours = 200 hours
Scenario B: Build tool to generate features
Time: 10 hours to build tool + 100 × 0.1 hours to generate
Time: 10 + 10 = 20 hours
Savings: 180 hours (90% reduction)
The same AI that builds features can build the infrastructure that builds features.
The Solution
Think like a Factorio player: don’t hand-craft items, build factories that produce items. Don’t build factories, build factories that build factories.
The Four Levels of Automation
Level 0: Manual Coding
You write all code yourself.
// You type every character
function authenticateUser(email: string, password: string) {
// 50 lines of authentication logic
}
Productivity: Baseline (1x)
Level 1: AI-Assisted Coding
You use Claude Code to write features.
Prompt: "Implement user authentication with JWT tokens"
Claude Code generates:
- authenticateUser function
- generateToken function
- verifyToken middleware
- Tests
Productivity: 5-10x (industry standard with AI coding agents)
Level 2: Building Tools with AI
You use Claude Code to build tools that generate code.
Prompt: "Build an MCP server that scaffolds CRUD endpoints"
Claude Code generates:
- MCP server implementation
- Template system for endpoints
- Test generator
- Documentation generator
Now you can:
mcp-scaffold create-crud User
→ Generates complete CRUD in 5 seconds
Productivity: 20-50x (automation multiplier)
Level 3: Meta-Infrastructure
You build tools that build tools.
Prompt: "Build a system that monitors our codebase, identifies
repetitive patterns, and auto-generates tools to eliminate them"
Claude Code generates:
- Pattern detection system
- Tool template generator
- Auto-deployment pipeline
- Usage monitoring
The system:
1. Detects you've written 5 similar API endpoints
2. Generates an endpoint generator tool
3. Refactors existing endpoints to use the tool
4. Documents the tool
5. Deploys it to your toolchain
Productivity: 100-500x (infrastructure compounds)
The Mindset Shift
Stop thinking: “How do I build this feature?”
Start thinking: “How do I build a system that builds features like this?”
Every time you encounter a task, ask:
- Is this a one-time task? → Use AI directly (Level 1)
- Will I do this 3+ times? → Build a tool (Level 2)
- Do I build tools like this often? → Build a tool generator (Level 3)
Implementation
Step 1: Identify High-Leverage Infrastructure
Not all infrastructure is equal. Focus on high-leverage projects that multiply capacity:
1. Custom MCP Servers
MCP servers extend Claude Code with custom capabilities.
Examples:
- CMS Integration Server: Fetch/update content from your CMS
- Performance Tracking Server: Monitor build times, bundle sizes, lighthouse scores
- Code Generation Server: Scaffold components, endpoints, tests
- Database Migration Server: Generate migrations from schema changes
Impact: Turn 20-minute tasks into 20-second commands
Build with Claude Code:
Prompt: "Build an MCP server that integrates with our headless CMS.
Should support:
- Fetching articles by ID/slug
- Creating/updating articles
- Uploading images
- Schema validation
Use the MCP SDK and follow the official examples."
2. Development Bootstrap Tools
Tools that initialize new projects with best practices baked in.
Examples:
- Project Scaffolder: Generate new projects with CI/CD, testing, linting pre-configured
- Feature Generator: Scaffold complete features (model + API + UI + tests)
- Integration Setup: One-command setup for databases, auth, monitoring
Impact: Reduce project setup from 2 days to 2 minutes
Build with Claude Code:
Prompt: "Build a CLI tool that scaffolds new Next.js projects with:
- TypeScript + tRPC + Prisma pre-configured
- Testing (Vitest + Playwright) set up
- CI/CD (GitHub Actions) configured
- CLAUDE.md files with our standards
- Git hooks for quality gates
Should support templates for: web-app, api-only, monorepo"
3. Evaluation Suite Infrastructure
Automated quality monitoring and optimization.
Examples:
- Visual Regression Testing: Auto-capture screenshots, detect UI changes
- Performance Monitoring: Track bundle size, lighthouse scores, load times
- Dependency Auditing: Monitor for vulnerabilities, outdated packages
- Quality Metrics Dashboard: Centralized view of test coverage, type safety, linting
Impact: Prevent regressions automatically instead of debugging them later
Build with Claude Code:
Prompt: "Build a visual regression testing system that:
1. Captures screenshots of all pages in development
2. Compares against baseline images
3. Flags differences >5% pixel change
4. Posts results to GitHub PR comments
5. Auto-updates baselines on approval
Use Playwright for screenshots, pixelmatch for comparison."
4. Claude Code Orchestration Systems
Automate Claude Code itself to work 24/7.
Examples:
- Nightly Job Orchestrator: Run tests, update dependencies, generate reports
- Ticket Distribution System: Assign GitHub issues to Claude Code automatically
- Progress Monitoring: Track completion, escalate blockers
- Auto-Review System: Claude Code reviews its own PRs against quality gates
Impact: Work continues while you sleep
Build with Claude Code:
Prompt: "Build a system that:
1. Monitors GitHub issues with label 'claude-task'
2. Spins up Claude Code instance for each issue
3. Tracks progress via issue comments
4. Creates PR when complete
5. Pings me if blocked for >1 hour
6. Automatically merges if all quality gates pass
Use GitHub API + Claude SDK + webhook listeners."
Step 2: Start with One High-Impact Tool
Don’t build everything at once. Start with the highest-leverage tool for your workflow.
How to identify highest-leverage:
-
Track your time for one week:
- What tasks take longest?
- What tasks are most repetitive?
- What tasks are most frustrating?
-
Calculate automation value:
Value = (Time per task) × (Frequency) × (Automation %) Example: Task: Setting up new API endpoints Time: 30 min per endpoint Frequency: 10 endpoints/week Automation: 90% (30 min → 3 min) Value = 30 × 10 × 0.9 = 270 min/week saved (4.5 hours) -
Estimate build time:
Build time: 4 hours to create endpoint scaffolder Payback: 4 hours / 4.5 hours per week = 0.9 weeks ROI: Positive after 5 days, then saves 4.5 hours/week forever -
Pick the tool with best ROI
Step 3: Build the Tool with Claude Code
Use Claude Code to build your infrastructure.
Template prompt:
I want to build [TOOL NAME] to automate [TASK].
Current process:
1. [Manual step 1]
2. [Manual step 2]
3. [Manual step 3]
Desired process:
1. Run: [command]
2. Done
Requirements:
- [Requirement 1]
- [Requirement 2]
- [Requirement 3]
Constraints:
- Use [language/framework]
- Follow [standards from CLAUDE.md]
- Include comprehensive tests
- Add error handling and logging
Deliverables:
1. Working implementation
2. Tests (>80% coverage)
3. Documentation (README + examples)
4. Integration with existing toolchain
Please:
1. Confirm you understand the requirements
2. Propose an architecture
3. Implement with tests
4. Create usage documentation
Step 4: Use the Tool and Iterate
After building your tool:
- Use it immediately on real tasks
- Track metrics:
- Time saved per use
- Frequency of use
- Bugs encountered
- Missing features
- Iterate:
- Fix bugs
- Add features based on usage
- Improve error messages
- Expand capabilities
Step 5: Build Your Second Tool
Once your first tool is providing value, build the next highest-leverage tool.
Compounding effect:
Tool 1 (Week 1): Saves 4 hours/week
→ Free time to build Tool 2
Tool 2 (Week 2): Saves 3 hours/week
→ Now saving 7 hours/week
→ Free time to build Tools 3, 4, 5
Tools 3-5 (Week 3-4): Save 10 hours/week
→ Now saving 17 hours/week
→ Free time to build meta-tools
Meta-tools (Week 5+): Auto-generate new tools
→ Exponential growth
Step 6: Build Tools That Build Tools
Once you have 5-10 tools, look for meta-patterns.
Example: You’ve built:
- API endpoint scaffolder
- Database migration generator
- React component generator
- Test file generator
Meta-pattern: All are “code generators from templates”
Meta-tool: Build a universal code generator framework:
// CLI: generate <template> <name> [options]
// Example: generate api-endpoint User --crud
interface Template {
name: string;
files: FileTemplate[];
hooks?: {
beforeGenerate?: () => void;
afterGenerate?: () => void;
};
}
interface FileTemplate {
path: string; // e.g., "src/api/{{name}}.ts"
content: string; // Template with {{variables}}
}
// Now adding new generators is just creating template files
// No code needed!
Prompt for Claude Code:
I've built several code generators and noticed they all follow
the same pattern: templates + variable substitution + file creation.
Build a universal code generator framework that:
1. Defines a template schema (JSON/YAML)
2. Supports variable substitution
3. Handles file creation with proper paths
4. Runs before/after hooks
5. Validates generated code
6. Provides CLI interface
This will let me create new generators by just writing templates,
no code needed.
Now creating new generators is 95% faster:
# templates/react-component.yaml
name: react-component
files:
- path: "src/components/{{name}}/{{name}}.tsx"
content: |
import React from 'react';
interface {{name}}Props {}
export function {{name}}({}: {{name}}Props) {
return <div>{{name}}</div>;
}
- path: "src/components/{{name}}/{{name}}.test.tsx"
content: |
import { render } from '@testing-library/react';
import { {{name}} } from './{{name}}';
describe('{{name}}', () => {
it('renders', () => {
render(<{{name}} />);
});
});
Run: generate react-component UserProfile
Done. Component + tests created.
Real-World Examples
Example 1: Custom MCP Server for Content Generation
Problem: Generating articles for compound.engineering is manual:
- Read IDEAS_TAXONOMY.md
- Extract idea details
- Craft prompt for Claude
- Generate article JSON
- Validate against schema
- Generate infographic
- Save to
/articles/
Time: ~30 minutes per article × 90 articles = 45 hours
Solution: Build an article generation MCP server
# Before: Manual process (30 min per article)
# After: Automated
mcp-article generate --idea semantic-naming-patterns
# 2 minutes per article
Savings: 28 minutes × 90 articles = 42 hours (93% reduction)
How it works:
- MCP server reads IDEAS_TAXONOMY.md
- Extracts idea by ID
- Reads schema and example articles
- Generates complete JSON
- Validates against schema
- Generates infographic with OpenAI
- Saves files
Implementation: See tools/article-generator/ in this repo
Example 2: Development Bootstrap CLI
Problem: Starting new projects requires:
- Clone template repo
- Update package.json
- Configure TypeScript
- Set up testing
- Configure linting
- Add CI/CD
- Create CLAUDE.md files
- Set up git hooks
- Initialize database
- Configure environment
Time: ~4 hours per project
Solution: One-command bootstrap
create-app my-project --template fullstack
# 3 minutes
Savings: 237 minutes per project
Generated project includes:
- Next.js + tRPC + Prisma configured
- Vitest + Playwright set up
- GitHub Actions CI/CD
- Pre-configured CLAUDE.md with standards
- Git hooks for linting/testing
- Docker compose for local development
- Environment templates
Build with Claude Code:
Prompt: "Build a CLI tool 'create-app' that scaffolds projects.
Support templates:
- fullstack: Next.js + tRPC + Prisma + Postgres
- api: Express + Prisma + Postgres
- frontend: Next.js only
Each template should include:
- All dependencies pre-configured
- Testing framework set up
- CI/CD pipeline configured
- CLAUDE.md files with our standards
- Git hooks for quality gates
- Docker compose for local development
- README with setup instructions
CLI should:
1. Ask for project name and template
2. Clone template from templates/ directory
3. Replace placeholders ({{PROJECT_NAME}}, etc.)
4. Initialize git repo
5. Install dependencies
6. Run initial tests to verify setup
7. Print next steps
Use @clack/prompts for interactive CLI."
Example 3: Nightly Job Orchestrator
Problem: Manual tasks that should run nightly:
- Update dependencies (check for updates, test, create PR)
- Run full test suite (including slow integration tests)
- Generate performance reports
- Check for security vulnerabilities
- Optimize images
- Update documentation
Time: 2 hours/night if done manually (never happens)
Solution: Nightly job orchestrator
// nightly-jobs.config.ts
export default {
schedule: '0 2 * * *', // 2 AM daily
jobs: [
{
name: 'update-dependencies',
command: 'npm update && npm test',
onSuccess: 'create-pr',
onFailure: 'notify-slack',
},
{
name: 'security-audit',
command: 'npm audit --audit-level=moderate',
onFailure: 'create-issue',
},
{
name: 'performance-report',
command: 'npm run perf-test',
onSuccess: 'save-report',
},
],
};
Build with Claude Code:
Prompt: "Build a nightly job orchestrator that:
1. Reads job configuration from nightly-jobs.config.ts
2. Runs jobs on schedule (using node-cron)
3. Supports success/failure hooks:
- create-pr: Create GitHub PR with changes
- create-issue: Create GitHub issue with details
- notify-slack: Send Slack notification
- save-report: Commit report to reports/ directory
4. Logs all job executions to logs/ directory
5. Provides dashboard at /nightly-jobs showing:
- Last run time
- Success/failure status
- Output logs
- Trend over time
6. Can be run manually via CLI: `npm run nightly-jobs now`
Use node-cron for scheduling, GitHub API for PR/issue creation."
Result: Maintenance happens automatically, overnight
Example 4: Tool-Building Tool (Meta-Infrastructure)
Problem: You’ve built 10 tools and realized they all follow similar patterns:
- Parse command-line arguments
- Validate inputs
- Load configuration
- Execute main logic
- Handle errors
- Log results
Solution: Build a “tool builder” that generates tools
create-tool scaffold-api-endpoint
# Generates:
# - CLI skeleton with argument parsing
# - Config file schema
# - Error handling
# - Logging setup
# - Tests
# - Documentation template
# You only need to fill in the main logic (20% of work)
Build with Claude Code:
Prompt: "Build a tool generator 'create-tool' that scaffolds
CLI tools following our standard patterns.
Generated tool should include:
1. CLI argument parsing (using commander.js)
2. Config file loading (from .config.ts)
3. Input validation (using zod)
4. Error handling (with proper exit codes)
5. Logging (using winston)
6. Progress indicators (using ora)
7. Tests (using vitest)
8. README template
The tool should have a main() function where the actual
logic goes - everything else is boilerplate.
Example usage:
create-tool my-scaffolder
# Generates tool skeleton
# Developer fills in main() function
# Tool is ready to publish
"
The Infrastructure Multiplication Effect
As you build more infrastructure, your capacity compounds:
Week 1: First Tool
Time investment: 8 hours
Time saved: 2 hours/day
Net: -8 hours (investment)
Week 2: Three More Tools
Time investment: 6 hours (faster due to Tool 1)
Time saved: 6 hours/day (cumulative)
Net: +42 hours for the week
Week 4: Ten Tools
Time investment: 4 hours (most work automated)
Time saved: ~8 hours/day
Net: All feature work is now automated
Month 2: Meta-Infrastructure
Tools are building tools
Manual work approaches zero
You focus on:
- Strategy
- Architecture
- Reviewing auto-generated PRs
- Building meta-meta-infrastructure
Month 6: Self-Improving Infrastructure
Systems that:
1. Monitor usage patterns
2. Identify inefficiencies
3. Generate improvement tickets
4. Auto-assign to Claude Code
5. Deploy updates automatically
You occasionally review and approve
Most work happens autonomously
Measuring Success
Key Metrics
1. Automation Coverage
Automation Coverage = (Automated Tasks) / (Total Recurring Tasks)
Target: >70% within 3 months
Track: Maintain a list of all recurring tasks and mark which are automated
2. Time Savings
Weekly Time Saved = Σ (Task Time × Frequency × Automation %)
Target: >20 hours/week within 2 months
Track: Before/after time logs for automated tasks
3. Tool ROI
ROI = (Time Saved) / (Build Time)
Target: >5x within 1 month of building tool
Track: Build time vs cumulative time saved
4. Infrastructure Leverage
Leverage = (Output with tools) / (Output without tools)
Target: >10x within 6 months
Track: Features shipped per week, before/after infrastructure
5. Tool Reuse
Reuse Rate = (Tools used by others) / (Total tools)
Target: >30% (tools useful beyond just you)
Track: Tool usage across team
Best Practices
1. Start Small, Think Big
Don’t try to automate everything at once.
✅ Good:
Week 1: Build one tool that saves 2 hours/day
Week 2: Build another tool
Week 3: Build meta-tool once pattern is clear
❌ Bad:
Week 1: Try to build complete automation framework
Week 2: Still building framework
Week 3: Abandoned due to scope creep
2. Build for Reuse
Every tool should be:
- Configurable: Use config files, not hardcoded values
- Composable: Can be used as part of larger workflows
- Documented: README + examples
- Tested: Automated tests so it doesn’t break
3. Optimize for Build Time
Use AI to build infrastructure fast:
Instead of spending 40 hours building perfect tool:
- Spend 4 hours with Claude Code building 80% solution
- Use it for 1 week
- Spend 4 hours improving based on usage
- Now you have 95% solution in 20% the time
4. Monitor Usage
Track which tools are actually used:
// Add usage tracking to all tools
import { trackUsage } from './analytics';
export async function main() {
trackUsage('tool-name', { action: 'run' });
// ... tool logic
}
// Weekly review:
// - Which tools are used most? → Invest more
// - Which tools are unused? → Deprecate
// - What patterns emerge? → Build meta-tools
5. Share Infrastructure
Tools that help you probably help others:
✅ Open source your infrastructure
✅ Document usage patterns
✅ Create onboarding guides
✅ Monitor adoption
Benefit: Others improve your tools
6. Build Feedback Loops
Infrastructure should improve itself:
// Example: Tool that monitors tool usage
export async function analyzeToolUsage() {
const usage = await getUsageMetrics();
// Identify inefficiencies
const inefficiencies = usage
.filter(tool => tool.avgDuration > 60_000) // >1 min
.map(tool => ({
tool: tool.name,
issue: 'Slow execution',
suggestion: 'Add caching or parallel execution',
}));
// Auto-create improvement tickets
for (const item of inefficiencies) {
await createGitHubIssue({
title: `Optimize ${item.tool}`,
body: `${item.issue}\n\nSuggestion: ${item.suggestion}`,
labels: ['infrastructure', 'performance', 'claude-task'],
});
}
}
// Run weekly via nightly job orchestrator
Common Pitfalls
❌ Pitfall 1: Building Before Validating Need
Problem: Building tools for tasks you rarely do
Solution: Track frequency before building. Only build if task happens >3 times.
❌ Pitfall 2: Over-Engineering
Problem: Spending 40 hours building perfect tool for 10-hour problem
Solution: Use “80% solution in 20% time” rule. Build fast, iterate.
❌ Pitfall 3: Building Without Using
Problem: Building tool but continuing to do task manually out of habit
Solution: Force yourself to use new tool immediately. Delete old manual process.
❌ Pitfall 4: Not Documenting
Problem: Building tool then forgetting how to use it 2 weeks later
Solution: Write README with examples BEFORE building. Claude Code can generate both.
❌ Pitfall 5: Single-Use Tools
Problem: Building tool that only works for one specific case
Solution: Add configuration. Turn generate-user-endpoint into generate-endpoint --model User.
Integration with Other Patterns
Combine with 24/7 Development Strategy
Building the Factory provides the infrastructure for continuous development:
- Build orchestration tools → Claude Code works overnight
- Build monitoring tools → Track progress while you sleep
- Build deployment tools → Auto-deploy when tests pass
See: 24/7 Development Strategy
Combine with Agent SDK Multi-Step Workflows
Building the Factory creates the tools that agents use:
- Build MCP servers → Agents can access your systems
- Build validation tools → Agents verify their work
- Build deployment tools → Agents ship features end-to-end
Combine with Quality Gates
Building the Factory automates quality enforcement:
- Build linting tools → Enforce code style
- Build testing tools → Catch regressions
- Build CI/CD tools → Block bad deployments
See: Quality Gates as Information Filters
The Ultimate Goal: Self-Improving Infrastructure
The endgame is infrastructure that improves itself:
┌─────────────────────────────────────────────────┐
│ Self-Improving Infrastructure │
└────────────────┬────────────────────────────────┘
│
▼
┌────────────────┐
│ Usage Monitor │ ← Track how tools are used
└────────┬───────┘
│
▼
┌───────────────────────┐
│ Inefficiency Detector │ ← Identify bottlenecks
└───────────┬───────────┘
│
▼
┌──────────────────────┐
│ Improvement Generator│ ← Generate tickets
└──────────┬───────────┘
│
▼
┌─────────────────────┐
│ Claude Code Agent │ ← Implement fixes
└─────────┬───────────┘
│
▼
┌────────────────────┐
│ Quality Gate Check │ ← Verify improvements
└────────┬───────────┘
│
▼
┌───────────────────┐
│ Auto-Deploy │ ← Ship improvements
└───────────────────┘
Implementation:
// monitor-and-improve.ts
export async function monitorAndImprove() {
// 1. Monitor usage
const metrics = await collectInfrastructureMetrics();
// 2. Detect inefficiencies
const issues = analyzeMetrics(metrics);
// 3. Generate improvement tickets
for (const issue of issues) {
await createGitHubIssue({
title: `Infrastructure: ${issue.title}`,
body: issue.description,
labels: ['infrastructure', 'auto-generated', 'claude-task'],
});
}
// 4. Claude Code picks up tickets and implements
// (via nightly job orchestrator)
// 5. Quality gates verify
// (via CI/CD)
// 6. Auto-deploy if tests pass
// (via deployment automation)
}
// Run daily
// Infrastructure improves itself continuously
Conclusion
Stop using AI to build features. Use AI to build infrastructure that builds features.
The levels of automation:
- Level 0: Manual coding (1x productivity)
- Level 1: AI-assisted coding (5-10x productivity) ← Most developers stop here
- Level 2: Building tools with AI (20-50x productivity)
- Level 3: Building tools that build tools (100-500x productivity)
The mindset shift:
- From: “How do I build this feature?”
- To: “How do I build a system that builds features like this?”
The result:
- Week 1: Linear productivity (use AI for tasks)
- Month 2: Exponential productivity (tools multiply capacity)
- Month 6: Compound productivity (infrastructure improves itself)
Start today:
- Identify your most repetitive task
- Calculate ROI of automating it
- Use Claude Code to build the tool
- Use the tool immediately
- Build your next tool
- Build tools that build tools
- Build systems that improve themselves
Think like a Factorio player: Don’t mine ore by hand. Build miners. Don’t place miners by hand. Build systems that place miners. Don’t optimize systems by hand. Build systems that optimize systems.
Build the factory.
Related Concepts
- 24/7 Development Strategy: Using infrastructure for continuous development
- Agent SDK Multi-Step Workflows: Giving agents the tools to work autonomously
- Quality Gates as Information Filters: Automating quality enforcement
- Hierarchical CLAUDE.md Files: Documentation infrastructure for AI context
- Custom ESLint Rules for Determinism: Building custom quality gates
Related Concepts
- 24/7 Development Strategy
- Quality Gates as Information Filters
- Hierarchical Context Patterns
- Custom ESLint Rules for Determinism
- Sub-Agent Architecture
- MCP Server for Dynamic Project Context – Build queryable knowledge infrastructure for AI agents
- Institutional Memory Learning Files – Meta-infrastructure for preserving decisions and rationales
References
- Factorio – The game that inspired this automation philosophy
- Model Context Protocol (MCP) – Build custom servers to extend AI coding agents
- Claude Agent SDK – Build autonomous agents with Claude

