multi-ai-implementation
Multi-AI Implementation
Overview
multi-ai-implementation provides systematic code generation and incremental development using the proven Explore-Plan-Code-Commit workflow with built-in TDD, quality gates, and multi-agent coordination.
Purpose: Transform specifications into production-ready code through incremental, test-driven development
Pattern: Workflow-based (6-step sequential process)
Key Principles (validated by tri-AI research):
- Explore Before Coding - Gather context first, never jump straight to implementation
- Plan Architecture - Use multi-ai-planning before writing code
- Incremental Development - Small changes (<200 lines), continuous testing
- Test-Driven - Write/generate tests first, then implement
- Quality Gates - Multi-layer validation before commit
- Safe Rollback - Automatic revert on test failures
Quality Guarantee: Code generated through this workflow achieves ≥85/100 quality score with ≥80% test coverage
When to Use
Use multi-ai-implementation when:
- Implementing new features (>200 lines, multiple components)
- Refactoring existing code (maintaining backward compatibility)
- Migrating systems (database, framework, architecture changes)
- Integrating components (API, services, third-party libraries)
- Building complex functionality requiring testing and verification
- Multi-agent coordination needed for parallel development
When NOT to Use:
- Simple changes (<50 lines, single file, obvious implementation)
- Documentation-only changes
- Configuration tweaks (use fast-track workflow instead)
Prerequisites
Required
- Clear implementation specification or plan
- Test framework available (Jest, Vitest, pytest, etc.)
- Time allocated (2-8 hours depending on complexity)
Recommended
- multi-ai-planning - Create plan before implementation
- multi-ai-testing - Generate and execute tests
- multi-ai-verification - Quality validation before commit
Understanding
- Basic TDD concepts (test-first development)
- Git workflow (commits, branches, rollback)
- Testing frameworks for your language
Implementation Workflow
Step 1: Explore & Gather Context
Gather comprehensive context through progressive disclosure before writing any code.
Purpose: Understand existing system, patterns, and constraints
Inputs:
- Implementation objective or specification
- Existing codebase (if applicable)
- Design mockups (for UI features)
- API documentation (for integrations)
Process:
-
Progressive File Discovery:
Use 3-level approach (never load everything):
Level 1: Structure (Glob)
# Map relevant files glob "**/*.{ts,js,py}" # Source files glob "**/test*.{ts,js,py}" # Test files glob "**/*auth*" # Domain-specificLevel 2: Patterns (Grep)
# Find specific patterns grep "export.*class.*Auth" --glob "**/*.ts" grep "describe\|it\(" --glob "**/*.test.ts" grep "TODO|FIXME" --glob "src/**"Level 3: Targeted Reading
# Read only critical files read "src/auth/login.ts" read "src/auth/tokens.ts" read "tests/auth/login.test.ts" -
Analyze Visual Mockups (for UI features):
- Read design files/screenshots
- Understand layout requirements
- Note component hierarchy
- Identify styling needs
-
Map Dependencies & Integration Points:
# Integration Analysis **Components This Touches**: - src/auth/* (authentication logic) - src/api/routes.ts (API endpoints) - src/middleware/auth.ts (authorization) **External Dependencies**: - jsonwebtoken (JWT generation) - bcrypt (password hashing) **Integration Points**: - Database (users, tokens tables) - API (POST /login, POST /register) - Frontend (auth forms, protected routes) -
Research Domain Patterns (optional - for unfamiliar domains):
If using multi-ai-research:
Use multi-ai-research for "Research [domain] implementation best practices and patterns"Manual research:
- Search: "[domain] best practices 2024-2025"
- Review: Official documentation
- Find: 3-5 example implementations
-
Document Context Synthesis:
# Context Synthesis ## Current State - [What exists now] - [Current patterns used] - [Existing tests] ## Desired State - [What should exist] - [New functionality needed] ## Constraints - Backward compatibility required - Must maintain existing API - Performance: <200ms response time ## Key Insights - [Pattern 1 found in codebase] - [Best practice from research] - [Integration consideration]
Outputs:
- Context synthesis document
- File inventory (relevant files identified)
- Pattern analysis
- Integration map
- Constraint documentation
Validation:
- Relevant files identified via progressive disclosure
- Existing patterns understood
- Integration points mapped
- Constraints documented
- Research conducted (if needed)
- Context comprehensive
Time Estimate: 30-60 minutes
Next: Proceed to Step 2
Step 2: Plan Architecture
Create detailed implementation plan using multi-ai-planning before writing any code.
Purpose: Think through architecture and approach before implementation
Inputs:
- Context synthesis (from Step 1)
- Implementation objective
- Success criteria
Process:
-
Invoke Multi-AI Planning:
Use multi-ai-planning to create implementation plan for [objective]This will guide you through:
- Objective analysis
- Hierarchical task decomposition
- Dependency mapping
- Verification planning
- Quality validation (≥90/100)
-
Review Generated Plan:
Plan Review Checklist: - [ ] All requirements covered by tasks - [ ] Tasks decomposed to atomic level - [ ] Dependencies mapped correctly - [ ] Each task has success criteria - [ ] Verification methods defined - [ ] Quality score ≥90/100 -
Extract Implementation Sequence:
# Implementation Sequence (from plan) Phase 1: Foundation (Tasks 1-2) - Task 1: Database schema - Task 2: Core models Phase 2: Business Logic (Tasks 3-4) - Task 3: Token management - Task 4: Authentication logic Phase 3: API Layer (Task 5) - Task 5: API endpoints Phase 4: Integration (Task 6) - Task 6: End-to-end integration -
Identify Test Strategy (from plan verification):
# Test Strategy **Test-First Approach**: - Generate tests before implementation - Confirm tests fail - Implement until tests pass **Coverage Targets**: - Gate (must pass): ≥80% line coverage - Target (desired): ≥95% line coverage **Test Types**: - Unit: Functions, classes - Integration: Components together - E2E: Complete workflows -
Optional: Use Extended Thinking for Complex Decisions:
- Press Tab for extended thinking mode
- Evaluate multiple architectural approaches
- Reason through trade-offs
- Document decision with rationale
Outputs:
- Complete implementation plan (from multi-ai-planning)
- plan.json (machine-readable)
- PLAN.md (human-readable)
- Implementation sequence
- Test strategy
Validation:
- Plan created using multi-ai-planning
- Plan quality score ≥90/100
- All tasks have verification defined
- Implementation sequence clear
- Test strategy documented
Time Estimate: 30-90 minutes (most time in multi-ai-planning)
Next: Proceed to Step 3
Step 3: Incremental Implementation (TDD)
Implement features incrementally using test-driven development with continuous validation.
Purpose: Build quality code through small, tested increments
Inputs:
- Implementation plan (from Step 2)
- Context synthesis (from Step 1)
- Test strategy
Process:
-
For Each Task in Plan, Follow TDD Cycle:
3.1. Generate Tests First (use multi-ai-testing):
Use multi-ai-testing TDD workflow for Task [X] Specification: - [What this task should do] - Success criteria: [from plan] - Edge cases: [boundary conditions]This generates tests and confirms they FAIL.
3.2. Implement to Pass Tests (incremental):
// Implement feature incrementally // Target: <200 lines per commit // Keep tests running continuously // Example: Implement token generation export function generateToken(user: User): string { // 1. Basic implementation (30 lines) const payload = { userId: user.id }; const token = jwt.sign(payload, process.env.JWT_SECRET); return token; } // Run tests → Some pass, some fail // 2. Add expiry (10 lines) // Run tests → More pass // 3. Add error handling (15 lines) // Run tests → All pass ✅3.3. Verify Tests Pass:
# Run test suite npm test -- src/auth/tokens.test.ts # Check coverage npm run coverage -- --file=src/auth/tokens.ts # Verify ≥80% coverage3.4. If Tests Fail: Iterate:
## Iteration Loop (Max 3 Attempts) Attempt 1: Implement feature → Tests fail ↓ Analyze failure, adjust code Attempt 2: Fix implementation → Tests fail (different error) ↓ Analyze, adjust again Attempt 3: Final fix → Tests pass ✅ If still failing after 3 attempts: ↓ HALT and request human review ❌ Doom Loop PreventionDoom Loop Breaker (Gemini recommendation):
- Max 3 implementation attempts per task
- If same error 3x: Escalate to human
- If oscillating errors (A→B→A): Escalate
- Prevents infinite refactoring loops
-
Maintain Incremental Commits:
Small Change Pattern:
# After each logical increment: git add <changed-files> git commit -m "Add [specific feature]: [what changed]" # Target: <200 lines per commit # Benefit: Clear intent, easy rollbackAutomatic Rollback on Test Failures:
# If tests fail after commit: npm test || git reset --hard HEAD~1 # TCR Pattern (Test-Commit-Revert): # - Tests pass → Commit # - Tests fail → Automatic revert -
Handle Backward Compatibility (for changes to existing code):
Expand-Migrate-Contract Pattern:
### Example: Add OAuth to Existing Auth **Phase 1: Expand** - Add new OAuth tables/functions - Keep existing password auth - Both work in parallel **Phase 2: Migrate** - Dual-write (password + OAuth) - Gradual user transition - Validate both paths work **Phase 3: Contract** - Deprecate password auth - Remove old code - OAuth only **Benefit**: Zero downtime, safe rollback at each phase -
Error Handling Pattern:
// Every function needs error handling export function generateToken(user: User): string { // Validate input if (!user || !user.id) { throw new Error('Invalid user: missing id'); } // Validate environment if (!process.env.JWT_SECRET) { throw new Error('JWT_SECRET not configured'); } try { // Implementation const token = jwt.sign({ userId: user.id }, process.env.JWT_SECRET); return token; } catch (error) { // Graceful handling console.error('Token generation failed:', error); throw new Error(`Failed to generate token: ${error.message}`); } }
Outputs:
- Working code (incremental commits)
- All tests passing
- Coverage ≥80% (gate), target ≥95%
- Error handling complete
- Backward compatible (if applicable)
Validation:
- All tests pass
- Coverage meets gate (≥80%)
- Changes <200 lines per commit (mostly)
- Error handling present
- Backward compatible verified
- No doom loops encountered
Time Estimate: 2-8 hours (varies by complexity)
Next: Proceed to Step 4 (if multi-agent needed) or Step 5
Step 4: Multi-Agent Coordination (Optional)
For complex features requiring parallel development, coordinate multiple implementation agents.
Purpose: Parallelize independent work to save time
When to Use:
- Feature has multiple independent components
- Different agents can work on different modules
- Clear separation of concerns possible
- Time optimization important
When to Skip:
- Simple linear implementation
- Strong coupling between components
- Single developer workflow
Inputs:
- Implementation plan with parallel groups identified
- Clear boundaries between components
Process:
-
Identify Parallel Opportunities (from plan):
## Parallel Groups (from multi-ai-planning) **Parallel Group 1** (after Task 1): - Task 2.1: Frontend authentication component - Task 2.2: Backend API endpoints - Task 2.3: Database migrations These can run simultaneously (independent). -
Spawn Parallel Implementation Agents (Task tool):
// Coordinate multiple agents via Task tool const implementations = await Promise.all([ task({ description: "Implement frontend auth component", prompt: `Create React authentication component. Specifications: - Login form with email/password - Form validation - API integration hooks - Error handling Write to: src/components/Auth/LoginForm.tsx Write tests to: src/components/Auth/LoginForm.test.tsx Success criteria: - Component renders correctly - Validation works - Tests pass - Coverage ≥80%` }), task({ description: "Implement backend API endpoints", prompt: `Create authentication API endpoints. Endpoints: - POST /api/auth/login - POST /api/auth/register - POST /api/auth/refresh Write to: src/api/auth.ts Write tests to: src/api/auth.test.ts Success criteria: - All endpoints functional - Tests pass - Coverage ≥80%` }), task({ description: "Create database migrations", prompt: `Create database schema for authentication. Tables: - users (id, email, password_hash, created_at) - sessions (id, user_id, token, expires_at) Write migrations to: migrations/001_auth_schema.sql Write tests to: migrations/001_auth_schema.test.ts` }) ]); // All three execute in parallel // Each has isolated context // Results returned when all complete -
Integrate Results:
# Integration After Parallel Execution **Check Each Component**: - [ ] Frontend component complete and tested - [ ] Backend API complete and tested - [ ] Database migrations complete and tested **Integration Steps**: 1. Verify all components built correctly 2. Connect frontend → API 3. Connect API → database 4. Run integration tests 5. Verify end-to-end workflow -
Handle Coordination Issues:
## Common Coordination Challenges **Issue**: Components don't integrate - **Cause**: Mismatched interfaces - **Fix**: Review specifications, align interfaces, re-implement **Issue**: Duplicate work - **Cause**: Overlapping agent responsibilities - **Fix**: Clearer task boundaries in plan **Issue**: Conflicting changes - **Cause**: Agents modifying same files - **Fix**: Use git worktrees for isolation
Outputs:
- Multiple components implemented in parallel
- All components tested independently
- Integration verified
- Time saved (20-40% faster than sequential)
Validation:
- All parallel tasks completed
- Each component tested independently
- Integration successful
- No conflicts or duplicate work
- Time savings achieved
Time Estimate: 2-6 hours (potentially 20-40% faster than sequential)
Next: Proceed to Step 5
Step 5: Integration & End-to-End Testing
Integrate all components and verify complete workflows function correctly.
Purpose: Ensure components work together, catch integration issues
Inputs:
- Implemented code (from Step 3-4)
- Integration test strategy (from plan)
Process:
-
Integration Testing:
Generate Integration Tests (use multi-ai-testing):
Use multi-ai-testing test generation workflow for integration tests Components: - [List all components to integrate] Workflows to Test: - User registration → database → email confirmation - User login → token generation → API access - Token refresh → validation → new token Write tests to: tests/integration/auth-workflows.test.ts -
Execute Integration Tests:
# Run integration test suite npm test -- tests/integration/ # Verify all workflows work # Fix any integration issues found -
End-to-End Testing:
# E2E Test Scenarios **Scenario 1: Complete Registration Flow** 1. User submits registration form 2. API validates and creates user 3. Database stores user record 4. Email confirmation sent 5. User confirms email 6. User can log in **Scenario 2: Complete Login Flow** [Similar detailed steps]Execute E2E Tests:
# Run E2E test suite npm run test:e2e # Or use multi-ai-testing Use multi-ai-testing for E2E testing of [workflows] -
Performance Validation:
# Run performance tests npm run test:performance # Check response times # Verify: <200ms for critical paths # Identify bottlenecks if any -
Backward Compatibility Check (for changes to existing code):
# Run full test suite (old + new tests) npm test # Verify: All existing tests still pass # Verify: No breaking changes # Verify: APIs backward compatible
Outputs:
- Integration tests passing
- E2E tests passing
- Performance validated
- Backward compatibility verified
- Integration issues resolved
Validation:
- Integration tests generated and passing
- E2E workflows verified
- Performance acceptable (<200ms or spec)
- No backward compatibility breaks
- All existing tests still pass
Time Estimate: 1-3 hours
Next: Proceed to Step 6
Step 6: Quality Verification & Commit
Run final verification before committing, ensuring all quality gates pass.
Purpose: Multi-layer quality assurance before production
Inputs:
- Complete implementation (from Steps 3-5)
- All tests passing
- Plan verification criteria
Process:
-
Run Multi-Layer Verification (use multi-ai-verification):
Use multi-ai-verification for complete quality check of [implementation]This runs all 5 verification layers:
- Layer 1: Rules-based (linting, types, schema) - 95% automated
- Layer 2: Functional (tests, coverage) - 60-80% automated
- Layer 3: Visual (if UI) - 30-50% automated
- Layer 4: Integration (E2E, system) - 20-30% automated
- Layer 5: Quality scoring (0-100) - LLM-as-judge
-
Review Verification Report:
# Verification Results ## Layer 1: Rules ✅ - Linting: PASS - Type checking: PASS - Schema validation: PASS ## Layer 2: Functional ✅ - Tests: 95/95 passing - Coverage: 87% (≥80% gate) - Examples: All working ## Layer 3: Visual ✅ (if applicable) - Screenshots match mockups - Responsive design works ## Layer 4: Integration ✅ - E2E tests pass - API integration verified ## Layer 5: Quality Score - **Total**: 92/100 ✅ (≥90 gate) - Correctness: 19/20 - Functionality: 19/20 - Quality: 18/20 - Integration: 18/20 - Security: 18/20 **Status**: ALL GATES PASS ✅ -
If Quality <90: Iterate:
## Gap Analysis (if score <90) **Issues Found**: 1. [Issue 1] - Priority: High - Fix: [Specific action] 2. [Issue 2] - Priority: Medium - Fix: [Action] **Action**: Apply fixes, re-verify **Target**: Reach ≥90/100 -
Create Git Commit (only if all gates pass):
# Stage changes git add <files> # Commit with clear message git commit -m "feat: Add user authentication with OAuth - Implemented JWT token generation and validation - Added login/register API endpoints - Created database schema and migrations - Tests: 95/95 passing, coverage 87% - Quality score: 92/100 Closes #123" -
Create Pull Request (for team workflows):
# Push branch git push -u origin feature/auth # Create PR gh pr create --title "Add user authentication" --body "$(cat <<EOF ## Summary Implements user authentication with OAuth support. ## Changes - Database schema for users and tokens - JWT token generation and validation - Login/register API endpoints - Frontend auth components ## Testing - Unit tests: 95/95 passing - Integration tests: 12/12 passing - Coverage: 87% (gate: ≥80%) - Quality score: 92/100 ## Verification - ✅ All 5 verification layers pass - ✅ Security scan: No vulnerabilities - ✅ Backward compatible - ✅ Performance: <150ms avg response Ready for review. EOF )" -
Document Implementation:
# Implementation Complete **What Was Built**: - [List of components] **Tests Added**: - [Test files and coverage] **Quality Metrics**: - Quality score: 92/100 - Test coverage: 87% - Performance: <150ms **Next Steps**: - Code review by team - Deployment to staging - Production release
Outputs:
- Production-ready code
- All tests passing
- Coverage ≥80% (gate), ideally ≥95%
- Quality score ≥90/100
- Git commit(s) created
- PR created (if team workflow)
- Documentation updated
Validation:
- All 5 verification layers pass
- Quality score ≥90/100
- Test coverage ≥80%
- Git commit created
- PR created (if applicable)
- Documentation complete
Time Estimate: 30-90 minutes
Result: Production-ready, tested, verified implementation
Integration with Other Skills
With multi-ai-planning (Step 2)
Usage: Create implementation plan before coding
Benefits:
- Structured approach (not ad-hoc)
- Quality ≥90 plans
- Clear task breakdown
- Verification built-in
With multi-ai-testing (Steps 3, 5)
Usage: TDD workflow, test generation, coverage validation
Benefits:
- Test-first development
- Independent verification prevents gaming
- 95% coverage achievable
- Self-healing tests
With multi-ai-verification (Step 6)
Usage: Multi-layer quality assurance before commit
Benefits:
- 5-layer verification (automated → LLM-as-judge)
- Quality scoring (0-100)
- All gates must pass
- Actionable feedback
Workflow Modes
Standard Mode (Full Quality)
Use For: Features, refactorings, security changes, integrations
Process: All 6 steps with all 5 verification layers Time: 5-15 hours Quality: Maximum (all gates, score ≥90)
Fast-Track Mode (Gemini Recommendation)
Use For: Typos, documentation, minor config tweaks
Process: Steps 1-3 + Layer 1-2 verification only Time: 30-90 minutes Quality: Essential checks only
Usage: Explicitly request "fast-track" or skip verification step
Best Practices
1. Always Explore First (Step 1)
Never jump straight to coding. Context prevents errors.
2. Always Plan (Step 2)
Use multi-ai-planning for quality ≥90 plans.
3. Test-Driven Development (Step 3)
Tests first, then implementation. Prevents overfitting.
4. Small Increments (<200 Lines)
Prevents LLM degradation, enables easy rollback.
5. Continuous Testing
Run tests after each change, not just at end.
6. Independent Verification (Step 6)
Use multi-ai-verification for unbiased quality check.
Common Mistakes
Mistake 1: Skipping Exploration
Problem: Coding without understanding existing patterns Fix: Always complete Step 1
Mistake 2: No Plan
Problem: Ad-hoc implementation, rework needed Fix: Always use multi-ai-planning (Step 2)
Mistake 3: Large Monolithic Changes
Problem: >500 lines, hard to review, LLM quality degrades Fix: Incremental commits (<200 lines)
Mistake 4: Implementation-First (Not Test-First)
Problem: Tests fit to code (gaming) Fix: Generate tests first in Step 3
Mistake 5: Skipping Verification
Problem: Low-quality code reaches production Fix: Always run multi-ai-verification (Step 6)
Appendix A: Task Tool Coordination Patterns
Pattern 1: Sequential Verification Agent
// Implementation complete
const implResult = await task({
description: "Implement authentication",
prompt: `Implement user authentication.
Write code to: src/auth/
Write tests to: tests/auth/
Write summary to: implementation-summary.json`
});
// Independent verification (separate agent)
const verification = await task({
description: "Verify implementation independently",
prompt: `Review implementation in src/auth/.
Do NOT read previous conversation.
Verify against success-criteria.json ONLY.
Check:
- All success criteria met
- Tests pass
- Code quality
- Security
Write report to: verification-report.md`
});
// Read verification results
const report = readFile('verification-report.md');
if (report.score >= 90) {
// Approved for commit
} else {
// Apply fixes from feedback
}
Pattern 2: Parallel Component Implementation
// Build multiple components in parallel
const [frontend, backend, database] = await Promise.all([
task({
description: "Build frontend component",
prompt: "Create React auth component. Write to: src/components/Auth/"
}),
task({
description: "Build backend API",
prompt: "Create auth API endpoints. Write to: src/api/auth.ts"
}),
task({
description: "Create database schema",
prompt: "Create auth schema. Write to: migrations/"
})
]);
// Integrate results after all complete
Pattern 3: Shared State via Files
// Agent 1: Research
await task({
description: "Research OAuth patterns",
prompt: "Research OAuth 2.0 implementation. Write findings to: research.md"
});
// Agent 2: Plan based on research
await task({
description: "Create implementation plan",
prompt: "Read research.md. Create detailed plan. Write to: plan.json"
});
// Agent 3: Implement based on plan
await task({
description: "Implement OAuth",
prompt: "Read plan.json. Implement OAuth. Write code to: src/oauth/"
});
// Pattern: Agent A → file → Agent B reads → next file
Appendix B: Doom Loop Prevention
Detection Patterns
Oscillating Errors:
Attempt 1: Fix error A → Test B fails
Attempt 2: Fix error B → Test A fails
Attempt 3: Fix error A → Test B fails again
↓
DOOM LOOP DETECTED → Escalate to human
Same Error Repeatedly:
Attempt 1: Error: "undefined userId"
Attempt 2: Error: "undefined userId" (same fix tried)
Attempt 3: Error: "undefined userId" (stuck)
↓
ESCALATE → Human needed
Breaker Implementation
MAX_RETRIES = 3
error_history = []
for attempt in range(MAX_RETRIES):
result = implement_and_test()
if result.success:
break
error_history.append(result.error)
# Detect doom loop
if attempt >= 2:
# Same error 3 times?
if error_history[0] == error_history[1] == error_history[2]:
escalate_to_human("Same error 3x:", error_history[0])
break
# Oscillating?
if error_history[0] == error_history[2] and error_history[1] != error_history[0]:
escalate_to_human("Oscillating errors:", error_history)
break
# Continue trying
Appendix C: Technical Foundation
Orchestration Runtime
- Development: Claude Code Task tool
- CI/CD: GitHub Actions
- Data Contracts: JSON schemas (in schemas/)
- Artifact Storage: File system (.implementations/)
Tooling Per Language
JavaScript/TypeScript:
- Linter: ESLint
- Type checker: TypeScript compiler (tsc)
- Test framework: Jest or Vitest
- Coverage: c8 or nyc
- SAST: Semgrep or eslint-plugin-security
Python:
- Linter: Pylint/Ruff
- Type checker: mypy
- Test framework: pytest
- Coverage: pytest-cov
- SAST: Bandit
Rollback Strategy
- Code: Git tags + git worktree (NOT git reset)
- Database: Migration down scripts
- Feature flags: Gradual rollout with kill switch
Cost/Latency Controls
- Budget: $50/month cap for LLM-as-judge
- Caching: Cache verification results for 24h
- Fast-path: Skip Layer 5 for changes <50 lines
Quick Reference
The 6-Step Workflow
| Step | Purpose | Time | Tools Used |
|---|---|---|---|
| 1 | Explore & Context | 30-60m | Glob, Grep, Read, Research |
| 2 | Plan Architecture | 30-90m | multi-ai-planning |
| 3 | Implement (TDD) | 2-8h | multi-ai-testing, Write, Edit |
| 4 | Coordinate (optional) | 2-6h | Task tool (parallel) |
| 5 | Integrate & E2E | 1-3h | multi-ai-testing (integration) |
| 6 | Verify & Commit | 30-90m | multi-ai-verification, Git |
Total: 5-15 hours (standard mode) or 30-90 min (fast-track)
Quality Metrics
- Test Coverage: ≥80% gate, ≥95% target
- Quality Score: ≥90/100 required
- Change Size: <200 lines per commit ideal
- Verification Layers: All 5 must pass
- Performance: Meets specification
multi-ai-implementation delivers production-ready code through systematic Explore-Plan-Code-Commit workflow with TDD, multi-agent coordination, and rigorous quality gates - validated by Claude + Gemini + Codex research.
For examples, see examples/. For coordination patterns, see Appendix A.