🏗️ How It Works
AI Task Manager transforms chaotic AI development into structured, manageable workflows through progressive refinement, context isolation, and validation gates.
The Core Problem
Traditional AI assistant “plan mode” features attempt to plan AND execute in a single continuous conversation, leading to:
Context Overload
[User Request + Planning + Task 1 + Task 2 + ... + Task N] → AI loses track
No Scope Control
- AI adds “nice-to-have” features not requested
- Plans expand beyond original requirements
- Difficult to maintain focus
Missing Validation
- Plans immediately proceed to execution
- No opportunity for human review
- Mistakes in planning compound during implementation
The AI Task Manager Solution
Three-Phase Progressive Refinement
AI Task Manager introduces mandatory human review gates between distinct phases:
flowchart TD
A[User Request] --> B[📝 Phase 1: Planning]
B --> C{Human Review}
C --> D[📋 Phase 2: Task Generation]
D --> E{Human Review}
E --> F[🚀 Phase 3: Execution]
F --> G[✅ Validated Output]
B --> B1[Requirements<br/>Clarification]
B --> B2[Architecture<br/>Decisions]
B --> B3[Risk<br/>Assessment]
D --> D1[Atomic Task<br/>Breakdown]
D --> D2[Dependency<br/>Mapping]
D --> D3[Skill<br/>Assignment]
F --> F1[Parallel<br/>Execution]
F --> F2[Quality<br/>Gates]
F --> F3[Progress<br/>Tracking]
style A fill:#ffebee
style C fill:#fff3e0
style E fill:#fff3e0
style G fill:#e8f5e8
style B fill:#e3f2fd
style D fill:#f3e5f5
style F fill:#e8f5e8
Phase 1: Strategic Planning
Input: User request with project requirements
Process:
- AI asks clarifying questions
- Documents requirements and constraints
- Proposes technical approach
- Identifies risks and success criteria
Output: Comprehensive plan document
Human Review Gate:
- Verify requirements captured accurately
- Remove unnecessary features (YAGNI enforcement)
- Adjust technical approach if needed
- Approve or request changes
Why This Matters:
- Catches scope creep before any code is written
- Ensures AI understands actual requirements
- Allows human expertise to guide technical decisions
Phase 2: Task Decomposition
Input: Approved plan document
Process:
- AI breaks plan into atomic tasks (1-2 skills each)
- Maps dependencies between tasks
- Assigns skills for specialized agent deployment
- Generates execution blueprint with phases
Output: Task documents + dependency graph + execution blueprint
Human Review Gate:
- Review all task acceptance criteria
- Remove tasks outside core scope
- Adjust dependencies if needed
- Verify parallelization opportunities
Why This Matters:
- Final opportunity to control scope
- Ensures tasks are right-sized (not too complex or too simple)
- Confirms implementation approach before execution
Phase 3: Phased Execution
Input: Approved tasks and execution blueprint
Process:
- Execute tasks phase by phase
- Run independent tasks in parallel within each phase
- Validate with POST_PHASE hook after each phase
- Commit results and move to next phase
Output: Working implementation with commits per phase
Validation Gates (Configurable via POST_PHASE hook):
- Linting passes
- Tests pass
- Coverage thresholds met
- Security scans pass
- Documentation updated
Why This Matters:
- Quality enforced incrementally, not at the end
- Easy to identify which phase introduced issues
- Git commits create natural rollback points
Design Principles
1. Context Isolation
Problem: AI assistants struggle with large context windows
Solution: Each phase operates with focused context
Phase 1 Context:
User Request + Clarifying Questions + Architecture Knowledge
Phase 2 Context:
Plan Document + Task Breakdown Rules + Dependency Logic
Phase 3 Context (Per Task):
Single Task + Dependencies + Acceptance Criteria + Project Context
Benefit: AI maintains focus and quality throughout implementation, even on complex projects
2. Progressive Disclosure
Principle: Provide information at the right time, not all at once
Implementation:
- Beginners: Start with basic workflow (create plan → generate tasks → execute)
- Intermediate: Add customization (hooks for quality gates)
- Advanced: Customize templates, create workflow patterns
Navigation Structure:
Getting Started → Core Concepts → Customization & Extension
Benefit: Users aren’t overwhelmed; they learn incrementally
3. Validation Gates (YAGNI Enforcement)
Problem: AI assistants tend to over-engineer and add unnecessary features
Solution: Multiple checkpoints that ask “Is this explicitly requested?”
Validation Points:
- PRE_PLAN Hook: Reminds AI to avoid gold-plating
- Plan Review: Human removes unrequested features
- Task Review: Human deletes unnecessary tasks
- POST_PHASE Hook: Quality gates ensure each phase meets standards
Benefit: Ships exactly what’s needed, not what “might be nice”
4. Skill-Based Decomposition
Problem: Generalist AI handling all aspects leads to diluted quality
Solution: Tasks tagged with 1-2 skills, specialized agents deployed per skill
Example:
# Task 1
skills: ["database", "schema-design"]
→ Database specialist agent
# Task 2
skills: ["react-components", "ui"]
→ Frontend specialist agent
# Task 3
skills: ["api-endpoints", "authentication"]
→ Backend specialist agent
Benefit:
- Deeper domain expertise per task
- Better code quality from specialized context
- Parallel execution of independent tasks
5. Human-in-the-Loop at Decision Points
Principle: AI generates options, humans make decisions
Decision Points:
- After planning (approve technical approach)
- After task generation (approve scope and breakdown)
- Between phases (verify quality before continuing)
Benefit: Combines AI speed with human judgment
Extensibility Architecture
AI Task Manager is built for customization through three extension points:
1. Hooks System
Seven lifecycle hooks inject custom logic:
PRE_PLAN → Planning guidance
PRE_PHASE → Phase preparation
PRE_TASK_ASSIGNMENT → Agent selection
POST_PLAN → Plan validation
POST_TASK_GENERATION_ALL → Task refinement
POST_PHASE → Quality gates
POST_ERROR_DETECTION → Error handling
Why Hooks Matter:
- Encode project-specific quality gates
- Integrate with existing tools (CI/CD, testing, deployment)
- Enforce organizational standards
- Adapt workflow to team needs
Example Use Cases:
- POST_PHASE: Run TypeScript compilation, test suite, coverage check
- PRE_PLAN: Include architecture guidelines from design docs
- POST_ERROR_DETECTION: Send Slack notifications on failures
See Customization Guide for detailed hook examples.
2. Templates System
Five customizable templates structure artifacts:
Plan Template: Captures requirements, architecture, risks, success criteria Task Template: Defines acceptance criteria, dependencies, implementation notes Blueprint Template: Organizes phases and parallelization Execution Summary: Documents results and learnings Fix Broken Tests: Enforces test integrity
Why Templates Matter:
- Consistent structure across all plans and tasks
- Easily add project-specific sections
- Team alignment on documentation standards
Example Customizations:
- Add security checklist to task template
- Include performance requirements section in plan template
- Add compliance considerations for regulated industries
See Customization Guide for template modification examples.
3. Configuration Files
TASK_MANAGER.md: Project context for AI assistants
- Tech stack and frameworks
- Coding standards
- Architecture decisions
- Links to documentation
Why Configuration Matters:
- AI assistants reference this for every task
- Ensures consistency across all implementations
- Documents decisions for team members
Value Proposition
Organized Execution
flowchart LR
A[Complex Project] --> B[AI Task Manager]
B --> C[📋 Structured Plans]
B --> D[⏰ Phased Execution]
B --> E[✅ Quality Gates]
B --> F[🔄 Consistent Process]
style A fill:#ffebee
style B fill:#fff3e0
style C fill:#e8f5e8
style D fill:#e8f5e8
style E fill:#e8f5e8
style F fill:#e8f5e8
Benefits:
- Clear Roadmap: Know exactly what will be built
- Predictable Timeline: Phase-based progress tracking
- Quality Assurance: Validation gates prevent rework
- Team Alignment: Shared plans and task documents
- Knowledge Capture: Plans document decisions and rationale
Time Savings
Eliminated:
- Back-and-forth clarifications (clarified upfront in planning)
- Rework from misunderstood requirements (plan review catches this)
- Debugging scope creep (removed during task review)
- Finding where AI went off-track (phase-based commits pinpoint issues)
Gained:
- Parallel execution of independent tasks
- Reusable plan patterns for similar projects
- Automated quality validation via hooks
Consistent Process
Every Project:
- Create plan
- Review and refine scope
- Generate tasks
- Review and adjust breakdown
- Execute with quality gates
- Archive for future reference
Benefits:
- Onboard new team members faster (standard workflow)
- Reduce decision fatigue (process is defined)
- Improve over time (refine hooks and templates)
Technical Foundation
AI Task Manager leverages research-backed principles:
Cognitive Load Theory: Phased approach reduces AI context overload, resulting in 25% improvement in task completion accuracy (research-validated)
Progressive Refinement: Multiple validation gates catch issues early, reducing rework by enforcing scope control at each phase
Mixture of Agents (MoA): Specialized agents for each task outperform generalist approaches—research shows collaborative multi-agent systems produce higher quality outputs
Skill-Based Decomposition: Maximum 2 skills per task ensures focused execution and enables parallel processing
Human-in-the-Loop: Strategic human input at decision points maximizes quality while maintaining AI execution speed
Next Steps
- Customization Guide: Learn to customize hooks, templates, and workflows
- Workflow Patterns: Advanced patterns for complex projects
- Basic Workflow: Day-to-day usage guide
- Features: Complete feature overview