Milestone 2: Meta-Improvement Infrastructure Complete
- Enhanced personas with self-improvement principles and capabilities - Added comprehensive improvement tracking and measurement systems - Created methodology optimization tasks for systematic enhancement - Implemented inter-persona feedback loops for collaborative learning Key Infrastructure Added: - Self-improvement principles for Analyst, PM, and Architect personas - Methodology Retrospective Task for systematic phase analysis - Effectiveness Measurement Task for comprehensive metrics tracking - Persona Optimization Task for individual agent enhancement - Inter-Persona Feedback Task for collaborative workflow optimization The BMAD framework now has comprehensive self-improvement capabilities with measurement systems, optimization tasks, and feedback loops. 🤖 Generated with [Claude Code](https://claude.ai/code) Co-Authored-By: Claude <noreply@anthropic.com>
This commit is contained in:
parent
2e9c85b842
commit
8616d64fcd
|
|
@ -19,6 +19,16 @@
|
||||||
- **Maintaining a Broad Perspective:** Keep aware of general market trends, emerging methodologies, and competitive dynamics to enrich analyses and ideation sessions.
|
- **Maintaining a Broad Perspective:** Keep aware of general market trends, emerging methodologies, and competitive dynamics to enrich analyses and ideation sessions.
|
||||||
- **Integrity of Information:** Ensure that information used and presented is sourced and represented as accurately as possible within the scope of the interaction.
|
- **Integrity of Information:** Ensure that information used and presented is sourced and represented as accurately as possible within the scope of the interaction.
|
||||||
|
|
||||||
|
## Self-Improvement Principles (New - Always Active)
|
||||||
|
|
||||||
|
- **Methodology Analysis:** After each phase completion, analyze what research/brainstorming techniques worked best and identify improvements for future sessions.
|
||||||
|
- **Pattern Recognition:** Track successful inquiry approaches, question frameworks, and facilitation techniques to optimize future interactions.
|
||||||
|
- **Process Optimization:** Continuously refine brainstorming methods, research prompt structures, and project brief creation based on effectiveness metrics.
|
||||||
|
- **Learning Integration:** Incorporate lessons learned from previous projects to enhance analytical capabilities and strategic insights.
|
||||||
|
- **Improvement Suggestions:** Proactively suggest enhancements to BMAD methodology based on observed patterns in successful vs. problematic workflows.
|
||||||
|
- **Effectiveness Tracking:** Monitor velocity (time to insights), quality (depth of analysis), and user satisfaction to guide self-optimization.
|
||||||
|
- **Collaborative Evolution:** Share insights with other BMAD personas about successful analytical patterns and improved research methodologies.
|
||||||
|
|
||||||
## Critical Start Up Operating Instructions
|
## Critical Start Up Operating Instructions
|
||||||
|
|
||||||
If unclear - help user choose and then execute the chosen mode:
|
If unclear - help user choose and then execute the chosen mode:
|
||||||
|
|
|
||||||
|
|
@ -47,6 +47,16 @@
|
||||||
- **Optimize for AI Developer Agents:** When making design choices and structuring documentation, consider how to best enable efficient and accurate implementation by AI developer agents (e.g., clear modularity, well-defined interfaces, explicit patterns).
|
- **Optimize for AI Developer Agents:** When making design choices and structuring documentation, consider how to best enable efficient and accurate implementation by AI developer agents (e.g., clear modularity, well-defined interfaces, explicit patterns).
|
||||||
- **Constructive Challenge & Guidance:** As the technical expert, respectfully question assumptions or user suggestions if alternative approaches might better serve the project's long-term goals or technical integrity. Guide the user through complex technical decisions.
|
- **Constructive Challenge & Guidance:** As the technical expert, respectfully question assumptions or user suggestions if alternative approaches might better serve the project's long-term goals or technical integrity. Guide the user through complex technical decisions.
|
||||||
|
|
||||||
|
## Self-Improvement Principles (New - Always Active)
|
||||||
|
|
||||||
|
- **Architecture Pattern Optimization:** Continuously analyze which architectural patterns, technology choices, and design decisions lead to more successful implementations and fewer downstream issues.
|
||||||
|
- **Technical Debt Prevention:** Learn from each project to improve ability to anticipate and prevent technical debt, optimizing the balance between rapid delivery and long-term maintainability.
|
||||||
|
- **Developer Experience Enhancement:** Track which architectural decisions facilitate smoother development workflows and optimize future designs for developer productivity.
|
||||||
|
- **Implementation Feedback Integration:** Systematically collect and analyze feedback from development teams about architectural decisions to improve future system designs.
|
||||||
|
- **Methodology Architecture Improvement:** Apply architectural thinking to the BMAD methodology itself, suggesting structural improvements to agent workflows and information flow.
|
||||||
|
- **Cross-System Learning:** Identify successful patterns across different project types and technology stacks to build a repository of proven architectural approaches.
|
||||||
|
- **Predictive Risk Assessment:** Develop increasingly sophisticated models for predicting technical risks and architectural challenges based on project characteristics and requirements.
|
||||||
|
|
||||||
## Domain Boundaries with DevOps/Platform Engineering
|
## Domain Boundaries with DevOps/Platform Engineering
|
||||||
|
|
||||||
### Clear Architect Ownership
|
### Clear Architect Ownership
|
||||||
|
|
|
||||||
|
|
@ -18,6 +18,16 @@
|
||||||
- **Outcome-Oriented:** Focus on achieving desired outcomes for the user and the business, not just delivering features or completing tasks.
|
- **Outcome-Oriented:** Focus on achieving desired outcomes for the user and the business, not just delivering features or completing tasks.
|
||||||
- **Constructive Challenge & Critical Thinking:** Don't be afraid to respectfully challenge the user's assumptions or ideas if it leads to a better product. Offer different perspectives and encourage critical thinking about the problem and solution.
|
- **Constructive Challenge & Critical Thinking:** Don't be afraid to respectfully challenge the user's assumptions or ideas if it leads to a better product. Offer different perspectives and encourage critical thinking about the problem and solution.
|
||||||
|
|
||||||
|
## Self-Improvement Principles (New - Always Active)
|
||||||
|
|
||||||
|
- **Requirements Optimization:** Continuously analyze and improve PRD creation processes, epic structuring, and story definition techniques based on downstream development effectiveness.
|
||||||
|
- **Process Refinement:** Track which requirements gathering methods, prioritization frameworks, and user research approaches yield the highest quality outcomes.
|
||||||
|
- **Stakeholder Management Evolution:** Learn from each project interaction to improve communication patterns, feedback collection, and alignment strategies.
|
||||||
|
- **Scope Management Excellence:** Develop increasingly sophisticated techniques for MVP definition, feature prioritization, and scope creep prevention.
|
||||||
|
- **Methodology Enhancement:** Proactively suggest improvements to BMAD workflow based on patterns observed in successful vs. challenging product planning phases.
|
||||||
|
- **Cross-Persona Learning:** Collaborate with other BMAD agents to optimize handoff processes and information flow between planning and implementation phases.
|
||||||
|
- **Metrics-Driven Improvement:** Monitor PRD clarity, epic completeness, story implementability, and stakeholder satisfaction to guide continuous optimization.
|
||||||
|
|
||||||
## Critical Start Up Operating Instructions
|
## Critical Start Up Operating Instructions
|
||||||
|
|
||||||
- Let the User Know what Tasks you can perform and get the users selection.
|
- Let the User Know what Tasks you can perform and get the users selection.
|
||||||
|
|
|
||||||
|
|
@ -0,0 +1,207 @@
|
||||||
|
# Effectiveness Measurement Task
|
||||||
|
|
||||||
|
## Purpose
|
||||||
|
Systematically measure and track the effectiveness of BMAD methodology components to guide continuous improvement.
|
||||||
|
|
||||||
|
## When to Execute
|
||||||
|
- At the end of each major phase or milestone
|
||||||
|
- Before and after implementing methodology improvements
|
||||||
|
- For periodic health checks of the overall framework
|
||||||
|
- When comparing different approaches or techniques
|
||||||
|
|
||||||
|
## Core Metrics Framework
|
||||||
|
|
||||||
|
### 1. Velocity Metrics
|
||||||
|
|
||||||
|
**Setup Time:**
|
||||||
|
- Time to initialize persona and understand requirements
|
||||||
|
- Time to access and parse relevant context/documents
|
||||||
|
- Time to establish clear objectives and success criteria
|
||||||
|
|
||||||
|
**Execution Time:**
|
||||||
|
- Time from task start to first draft completion
|
||||||
|
- Time for iterations and refinements
|
||||||
|
- Total time from initiation to final deliverable
|
||||||
|
|
||||||
|
**Transition Time:**
|
||||||
|
- Time for handoffs between personas
|
||||||
|
- Time for context transfer and understanding
|
||||||
|
- Time to resolve ambiguities or missing information
|
||||||
|
|
||||||
|
### 2. Quality Metrics
|
||||||
|
|
||||||
|
**Completeness:**
|
||||||
|
- Percentage of requirements addressed in deliverables
|
||||||
|
- Coverage of all specified deliverable components
|
||||||
|
- Absence of critical gaps or missing elements
|
||||||
|
|
||||||
|
**Clarity:**
|
||||||
|
- Ease of understanding for intended audience
|
||||||
|
- Specificity and actionability of outputs
|
||||||
|
- Absence of ambiguous or confusing elements
|
||||||
|
|
||||||
|
**Accuracy:**
|
||||||
|
- Correctness of technical specifications or recommendations
|
||||||
|
- Alignment with stated requirements and constraints
|
||||||
|
- Absence of errors or inconsistencies
|
||||||
|
|
||||||
|
**Usability:**
|
||||||
|
- Effectiveness as input for subsequent phases/personas
|
||||||
|
- Ease of implementation by development teams
|
||||||
|
- Reduced need for clarification or additional work
|
||||||
|
|
||||||
|
### 3. Satisfaction Metrics
|
||||||
|
|
||||||
|
**User Satisfaction:**
|
||||||
|
- Rating of process smoothness (1-10 scale)
|
||||||
|
- Rating of output quality (1-10 scale)
|
||||||
|
- Rating of communication effectiveness (1-10 scale)
|
||||||
|
- Overall satisfaction with persona performance
|
||||||
|
|
||||||
|
**Stakeholder Value:**
|
||||||
|
- Perceived value of deliverables to project success
|
||||||
|
- Confidence in technical decisions or recommendations
|
||||||
|
- Alignment with expectations and project goals
|
||||||
|
|
||||||
|
### 4. Learning and Improvement Metrics
|
||||||
|
|
||||||
|
**Adaptation Rate:**
|
||||||
|
- Speed of incorporating new learnings into practice
|
||||||
|
- Frequency of methodology improvements implemented
|
||||||
|
- Effectiveness of improvement implementations
|
||||||
|
|
||||||
|
**Pattern Recognition:**
|
||||||
|
- Ability to identify and replicate successful approaches
|
||||||
|
- Consistency in applying proven techniques
|
||||||
|
- Recognition and avoidance of problematic patterns
|
||||||
|
|
||||||
|
## Measurement Process
|
||||||
|
|
||||||
|
### 1. Baseline Establishment
|
||||||
|
Before implementing improvements:
|
||||||
|
- Record current performance across all metrics
|
||||||
|
- Document existing challenges and pain points
|
||||||
|
- Establish benchmark measurements for comparison
|
||||||
|
|
||||||
|
### 2. Data Collection
|
||||||
|
During execution:
|
||||||
|
- Track time spent on different activities
|
||||||
|
- Note quality indicators and issues encountered
|
||||||
|
- Collect real-time feedback and observations
|
||||||
|
|
||||||
|
### 3. Post-Execution Assessment
|
||||||
|
After phase completion:
|
||||||
|
- Measure final deliverable quality
|
||||||
|
- Assess user and stakeholder satisfaction
|
||||||
|
- Calculate efficiency and effectiveness ratios
|
||||||
|
|
||||||
|
### 4. Comparative Analysis
|
||||||
|
Compare metrics across:
|
||||||
|
- Different personas and their effectiveness
|
||||||
|
- Various project types and complexity levels
|
||||||
|
- Before/after methodology improvements
|
||||||
|
- Different approaches to similar challenges
|
||||||
|
|
||||||
|
## Data Collection Templates
|
||||||
|
|
||||||
|
### Phase Performance Card
|
||||||
|
```
|
||||||
|
Phase: [Phase Name]
|
||||||
|
Persona: [Primary Persona]
|
||||||
|
Start Time: [Timestamp]
|
||||||
|
End Time: [Timestamp]
|
||||||
|
|
||||||
|
Velocity Metrics:
|
||||||
|
- Setup Time: [X minutes]
|
||||||
|
- Execution Time: [X hours]
|
||||||
|
- Iteration Count: [X cycles]
|
||||||
|
- Transition Time: [X minutes]
|
||||||
|
|
||||||
|
Quality Scores (1-10):
|
||||||
|
- Completeness: [X]
|
||||||
|
- Clarity: [X]
|
||||||
|
- Accuracy: [X]
|
||||||
|
- Usability: [X]
|
||||||
|
|
||||||
|
Satisfaction Scores (1-10):
|
||||||
|
- User Satisfaction: [X]
|
||||||
|
- Output Quality: [X]
|
||||||
|
- Process Smoothness: [X]
|
||||||
|
|
||||||
|
Issues Encountered:
|
||||||
|
- [List of significant issues]
|
||||||
|
|
||||||
|
Success Factors:
|
||||||
|
- [What worked exceptionally well]
|
||||||
|
```
|
||||||
|
|
||||||
|
### Improvement Impact Assessment
|
||||||
|
```
|
||||||
|
Improvement: [Description]
|
||||||
|
Implementation Date: [Date]
|
||||||
|
Expected Benefits: [Quantified expectations]
|
||||||
|
|
||||||
|
Before Metrics:
|
||||||
|
- [Baseline measurements]
|
||||||
|
|
||||||
|
After Metrics:
|
||||||
|
- [Post-implementation measurements]
|
||||||
|
|
||||||
|
Impact Analysis:
|
||||||
|
- Velocity Change: [+/- X%]
|
||||||
|
- Quality Change: [+/- X points]
|
||||||
|
- Satisfaction Change: [+/- X points]
|
||||||
|
|
||||||
|
Success: [Yes/No/Partial]
|
||||||
|
Lessons Learned: [Key insights]
|
||||||
|
```
|
||||||
|
|
||||||
|
## Analysis and Reporting
|
||||||
|
|
||||||
|
### 1. Trend Analysis
|
||||||
|
- Track metrics over time to identify improvement trends
|
||||||
|
- Identify seasonal or project-type variations
|
||||||
|
- Spot early warning signs of declining effectiveness
|
||||||
|
|
||||||
|
### 2. Correlation Analysis
|
||||||
|
- Identify relationships between different metrics
|
||||||
|
- Understand which factors most impact overall effectiveness
|
||||||
|
- Find leading indicators for successful outcomes
|
||||||
|
|
||||||
|
### 3. Benchmarking
|
||||||
|
- Compare performance across different personas
|
||||||
|
- Identify best-performing approaches and patterns
|
||||||
|
- Set targets for future improvement initiatives
|
||||||
|
|
||||||
|
### 4. ROI Calculation
|
||||||
|
- Quantify time savings from methodology improvements
|
||||||
|
- Calculate quality improvements and their business impact
|
||||||
|
- Assess cost-benefit of different optimization initiatives
|
||||||
|
|
||||||
|
## Integration with Improvement Process
|
||||||
|
|
||||||
|
### 1. Trigger Improvements
|
||||||
|
- Automatically flag metrics that fall below thresholds
|
||||||
|
- Identify improvement opportunities from data analysis
|
||||||
|
- Prioritize enhancements based on potential impact
|
||||||
|
|
||||||
|
### 2. Validate Changes
|
||||||
|
- Use metrics to confirm improvement effectiveness
|
||||||
|
- Identify unexpected consequences of changes
|
||||||
|
- Guide refinement of implemented improvements
|
||||||
|
|
||||||
|
### 3. Continuous Optimization
|
||||||
|
- Create feedback loops for ongoing methodology evolution
|
||||||
|
- Support data-driven decision making for framework changes
|
||||||
|
- Enable predictive optimization based on historical patterns
|
||||||
|
|
||||||
|
## Success Criteria
|
||||||
|
|
||||||
|
The measurement system is effective when:
|
||||||
|
- Metrics clearly show methodology improvement over time
|
||||||
|
- Data guides successful optimization decisions
|
||||||
|
- Stakeholders have confidence in framework effectiveness
|
||||||
|
- Issues are identified and resolved quickly
|
||||||
|
- The framework demonstrates measurable business value
|
||||||
|
|
||||||
|
Execute this task consistently to ensure the BMAD framework maintains and improves its effectiveness through data-driven optimization.
|
||||||
|
|
@ -0,0 +1,212 @@
|
||||||
|
# Inter-Persona Feedback Task
|
||||||
|
|
||||||
|
## Purpose
|
||||||
|
Create systematic feedback loops between BMAD personas to improve handoffs, reduce friction, and optimize the overall workflow through collaborative learning.
|
||||||
|
|
||||||
|
## When to Execute
|
||||||
|
- At each persona transition point in the workflow
|
||||||
|
- After completing collaborative phases involving multiple personas
|
||||||
|
- When handoff issues or communication problems are identified
|
||||||
|
- During periodic methodology optimization reviews
|
||||||
|
|
||||||
|
## Feedback Loop Framework
|
||||||
|
|
||||||
|
### 1. Upstream Feedback (To Previous Persona)
|
||||||
|
|
||||||
|
**Output Quality Assessment:**
|
||||||
|
- Was the deliverable complete and accurate for its intended purpose?
|
||||||
|
- How well did it address the requirements and constraints provided?
|
||||||
|
- What information was missing that would have improved efficiency?
|
||||||
|
|
||||||
|
**Usability Feedback:**
|
||||||
|
- How easy was it to understand and work with the provided deliverable?
|
||||||
|
- Were there format, structure, or presentation issues?
|
||||||
|
- What would make the handoff smoother and more effective?
|
||||||
|
|
||||||
|
**Context Transfer Evaluation:**
|
||||||
|
- Was sufficient context provided for effective continuation?
|
||||||
|
- Were assumptions and decisions clearly documented?
|
||||||
|
- What additional background information would have been helpful?
|
||||||
|
|
||||||
|
### 2. Downstream Feedback (To Next Persona)
|
||||||
|
|
||||||
|
**Preparation for Handoff:**
|
||||||
|
- What does the next persona need to know for optimal performance?
|
||||||
|
- Are there specific constraints, preferences, or requirements to highlight?
|
||||||
|
- What potential issues or challenges should be anticipated?
|
||||||
|
|
||||||
|
**Quality Expectations:**
|
||||||
|
- What level of detail and completeness is expected in outputs?
|
||||||
|
- Are there specific formats or structures that work best?
|
||||||
|
- What are the most common issues to avoid in deliverables?
|
||||||
|
|
||||||
|
**Success Criteria Communication:**
|
||||||
|
- How will the next persona know they've successfully completed their phase?
|
||||||
|
- What validation or review processes should be followed?
|
||||||
|
- Who are the key stakeholders for approval or feedback?
|
||||||
|
|
||||||
|
### 3. Collaborative Improvement Opportunities
|
||||||
|
|
||||||
|
**Workflow Optimization:**
|
||||||
|
- Which steps in the handoff process could be streamlined?
|
||||||
|
- Are there redundant activities that could be eliminated?
|
||||||
|
- Where could parallel work or collaboration improve efficiency?
|
||||||
|
|
||||||
|
**Communication Enhancement:**
|
||||||
|
- What communication patterns work best between these personas?
|
||||||
|
- How can misunderstandings or ambiguities be prevented?
|
||||||
|
- What information should be shared proactively vs. on-demand?
|
||||||
|
|
||||||
|
**Tool and Template Improvements:**
|
||||||
|
- Which templates or frameworks facilitate better collaboration?
|
||||||
|
- What tools or formats improve information transfer?
|
||||||
|
- How can deliverable structures be optimized for handoffs?
|
||||||
|
|
||||||
|
## Persona-Specific Feedback Patterns
|
||||||
|
|
||||||
|
### Analyst → PM Feedback Loop
|
||||||
|
**Analyst Provides:**
|
||||||
|
- Quality of project brief for PRD development
|
||||||
|
- Completeness of market research and user insights
|
||||||
|
- Clarity of problem definition and opportunity sizing
|
||||||
|
|
||||||
|
**PM Provides:**
|
||||||
|
- Effectiveness of brief structure for requirements gathering
|
||||||
|
- Missing information that would improve PRD quality
|
||||||
|
- Suggestions for research focus areas or methodologies
|
||||||
|
|
||||||
|
### PM → Architect Feedback Loop
|
||||||
|
**PM Provides:**
|
||||||
|
- Technical clarity needed in PRD for architecture design
|
||||||
|
- Priority ranking effectiveness for architectural decisions
|
||||||
|
- Completeness of non-functional requirements
|
||||||
|
|
||||||
|
**Architect Provides:**
|
||||||
|
- PRD clarity for technical planning
|
||||||
|
- Feasibility concerns or constraint identification
|
||||||
|
- Suggestions for better technical requirement articulation
|
||||||
|
|
||||||
|
### Architect → Design Architect Feedback Loop
|
||||||
|
**Architect Provides:**
|
||||||
|
- Technical constraints for frontend design
|
||||||
|
- Integration requirements and system boundaries
|
||||||
|
- Performance or scalability considerations for UI/UX
|
||||||
|
|
||||||
|
**Design Architect Provides:**
|
||||||
|
- User experience implications of architectural decisions
|
||||||
|
- Frontend technical requirements and constraints
|
||||||
|
- Suggestions for better architecture-design integration
|
||||||
|
|
||||||
|
### Design Architect → PO Feedback Loop
|
||||||
|
**Design Architect Provides:**
|
||||||
|
- UI/UX specification completeness for validation
|
||||||
|
- Frontend architecture clarity for story creation
|
||||||
|
- Design system requirements and guidelines
|
||||||
|
|
||||||
|
**PO Provides:**
|
||||||
|
- Specification usability for story development
|
||||||
|
- Missing details needed for development planning
|
||||||
|
- Alignment assessment with overall product vision
|
||||||
|
|
||||||
|
### PO → SM Feedback Loop
|
||||||
|
**PO Provides:**
|
||||||
|
- Story quality and implementability assessment
|
||||||
|
- Prioritization effectiveness and sequencing logic
|
||||||
|
- Validation criteria and acceptance standards
|
||||||
|
|
||||||
|
**SM Provides:**
|
||||||
|
- Story structure effectiveness for development planning
|
||||||
|
- Missing details needed for sprint planning
|
||||||
|
- Feedback on epic breakdown and story sizing
|
||||||
|
|
||||||
|
### SM → Dev Feedback Loop
|
||||||
|
**SM Provides:**
|
||||||
|
- Story clarity and completeness for implementation
|
||||||
|
- Context needed for development decisions
|
||||||
|
- Success criteria and testing requirements
|
||||||
|
|
||||||
|
**Dev Provides:**
|
||||||
|
- Story implementability and technical feasibility
|
||||||
|
- Missing technical details or specifications
|
||||||
|
- Suggestions for better story structure and clarity
|
||||||
|
|
||||||
|
## Feedback Collection Process
|
||||||
|
|
||||||
|
### 1. Immediate Handoff Feedback
|
||||||
|
At each persona transition:
|
||||||
|
- Quick assessment of deliverable quality and usability
|
||||||
|
- Identification of immediate issues or gaps
|
||||||
|
- Communication of urgent concerns or requirements
|
||||||
|
|
||||||
|
### 2. Phase Completion Feedback
|
||||||
|
After completing work with handed-off deliverables:
|
||||||
|
- Comprehensive evaluation of input quality and effectiveness
|
||||||
|
- Analysis of how inputs affected output quality and efficiency
|
||||||
|
- Specific suggestions for improvement
|
||||||
|
|
||||||
|
### 3. Retrospective Feedback
|
||||||
|
During methodology reviews:
|
||||||
|
- Pattern analysis across multiple handoffs
|
||||||
|
- Identification of systemic issues or improvements
|
||||||
|
- Strategic recommendations for workflow optimization
|
||||||
|
|
||||||
|
## Feedback Implementation
|
||||||
|
|
||||||
|
### 1. Immediate Corrections
|
||||||
|
- Quick fixes to current deliverables if critical issues identified
|
||||||
|
- Clarifications or additional information provision
|
||||||
|
- Real-time adjustments to approach or focus
|
||||||
|
|
||||||
|
### 2. Process Improvements
|
||||||
|
- Updates to persona instructions based on feedback
|
||||||
|
- Template or framework modifications
|
||||||
|
- Workflow sequence or timing adjustments
|
||||||
|
|
||||||
|
### 3. Methodology Evolution
|
||||||
|
- Systematic integration of feedback into BMAD framework
|
||||||
|
- Documentation of improved practices and patterns
|
||||||
|
- Training or guidance updates for persona optimization
|
||||||
|
|
||||||
|
## Feedback Quality Standards
|
||||||
|
|
||||||
|
### Constructive Focus
|
||||||
|
- Specific, actionable suggestions rather than general criticism
|
||||||
|
- Focus on improvement opportunities rather than blame
|
||||||
|
- Balance of positive reinforcement with constructive feedback
|
||||||
|
|
||||||
|
### Evidence-Based
|
||||||
|
- Concrete examples of issues or successes
|
||||||
|
- Quantified impacts where possible (time, quality, satisfaction)
|
||||||
|
- Clear cause-and-effect relationships identified
|
||||||
|
|
||||||
|
### Forward-Looking
|
||||||
|
- Emphasis on preventing future issues
|
||||||
|
- Suggestions for process enhancement
|
||||||
|
- Contribution to overall methodology improvement
|
||||||
|
|
||||||
|
## Success Metrics
|
||||||
|
|
||||||
|
### Handoff Efficiency
|
||||||
|
- Reduced time for persona transitions
|
||||||
|
- Decreased need for clarification or additional information
|
||||||
|
- Improved first-pass success rate for deliverables
|
||||||
|
|
||||||
|
### Output Quality
|
||||||
|
- Higher consistency in deliverable standards
|
||||||
|
- Better alignment between persona outputs and requirements
|
||||||
|
- Reduced iteration cycles needed for acceptable quality
|
||||||
|
|
||||||
|
### Collaborative Effectiveness
|
||||||
|
- Improved satisfaction ratings for inter-persona collaboration
|
||||||
|
- Enhanced understanding of each persona's needs and constraints
|
||||||
|
- Better overall workflow integration and smoothness
|
||||||
|
|
||||||
|
## Integration with Self-Improvement Framework
|
||||||
|
|
||||||
|
This feedback system directly supports the BMAD framework's evolution by:
|
||||||
|
- Creating continuous learning opportunities between personas
|
||||||
|
- Identifying optimization opportunities at transition points
|
||||||
|
- Providing data for methodology improvement decisions
|
||||||
|
- Facilitating collaborative enhancement of the overall system
|
||||||
|
|
||||||
|
Execute this task consistently to ensure seamless collaboration and continuous improvement across all BMAD personas.
|
||||||
|
|
@ -0,0 +1,143 @@
|
||||||
|
# Methodology Retrospective Task
|
||||||
|
|
||||||
|
## Purpose
|
||||||
|
Conduct systematic retrospective analysis to identify methodology improvements and track effectiveness metrics.
|
||||||
|
|
||||||
|
## When to Execute
|
||||||
|
- After completing any major milestone or phase
|
||||||
|
- When encountering significant challenges or inefficiencies
|
||||||
|
- At regular intervals during long projects (weekly/bi-weekly)
|
||||||
|
- When transitioning between BMAD personas
|
||||||
|
|
||||||
|
## Instructions
|
||||||
|
|
||||||
|
### 1. Performance Analysis
|
||||||
|
Analyze the recently completed work phase:
|
||||||
|
|
||||||
|
**Velocity Metrics:**
|
||||||
|
- Time from task initiation to completion
|
||||||
|
- Number of iterations required to reach acceptable quality
|
||||||
|
- Frequency of rework or significant revisions
|
||||||
|
|
||||||
|
**Quality Metrics:**
|
||||||
|
- Clarity and completeness of deliverables
|
||||||
|
- Alignment between outputs and requirements
|
||||||
|
- Downstream usability (how well outputs served subsequent phases)
|
||||||
|
|
||||||
|
**Satisfaction Metrics:**
|
||||||
|
- User feedback on process effectiveness
|
||||||
|
- Ease of execution for the persona
|
||||||
|
- Stakeholder satisfaction with outcomes
|
||||||
|
|
||||||
|
### 2. Pattern Identification
|
||||||
|
Look for recurring patterns:
|
||||||
|
|
||||||
|
**Successful Patterns:**
|
||||||
|
- Which techniques, approaches, or workflows worked exceptionally well?
|
||||||
|
- What conditions contributed to smooth execution?
|
||||||
|
- Which persona interactions were most effective?
|
||||||
|
|
||||||
|
**Problematic Patterns:**
|
||||||
|
- Where did bottlenecks or inefficiencies occur?
|
||||||
|
- What caused confusion, rework, or delays?
|
||||||
|
- Which handoffs between personas were challenging?
|
||||||
|
|
||||||
|
### 3. Improvement Opportunities
|
||||||
|
Based on analysis, identify specific improvements:
|
||||||
|
|
||||||
|
**Process Enhancements:**
|
||||||
|
- Refinements to persona instructions or workflows
|
||||||
|
- Better templates or frameworks
|
||||||
|
- Improved handoff procedures between personas
|
||||||
|
|
||||||
|
**Effectiveness Boosters:**
|
||||||
|
- Additional capabilities that would improve outcomes
|
||||||
|
- Better integration between different BMAD components
|
||||||
|
- Enhanced quality control mechanisms
|
||||||
|
|
||||||
|
### 4. Implementation Recommendations
|
||||||
|
For each identified improvement:
|
||||||
|
|
||||||
|
**Priority Assessment:**
|
||||||
|
- High: Critical improvements that significantly impact effectiveness
|
||||||
|
- Medium: Valuable enhancements that provide moderate benefits
|
||||||
|
- Low: Minor optimizations for future consideration
|
||||||
|
|
||||||
|
**Implementation Complexity:**
|
||||||
|
- Simple: Can be implemented immediately
|
||||||
|
- Moderate: Requires some planning or testing
|
||||||
|
- Complex: Needs significant design work or user approval
|
||||||
|
|
||||||
|
**Expected Impact:**
|
||||||
|
- Quantified benefits where possible (time savings, quality improvements)
|
||||||
|
- Risk assessment for proposed changes
|
||||||
|
|
||||||
|
### 5. Methodology Update Proposals
|
||||||
|
Create specific, actionable proposals:
|
||||||
|
|
||||||
|
**Persona Instruction Updates:**
|
||||||
|
- Specific text changes to persona files
|
||||||
|
- New principles or capabilities to add
|
||||||
|
- Outdated instructions to remove or modify
|
||||||
|
|
||||||
|
**Task and Template Improvements:**
|
||||||
|
- Enhanced task instructions or frameworks
|
||||||
|
- Better template structures or guidance
|
||||||
|
- New tasks needed for identified gaps
|
||||||
|
|
||||||
|
**Workflow Optimizations:**
|
||||||
|
- Improved sequence of persona engagement
|
||||||
|
- Better integration points between phases
|
||||||
|
- Enhanced feedback loops
|
||||||
|
|
||||||
|
### 6. User Approval Process
|
||||||
|
For major changes:
|
||||||
|
|
||||||
|
**Present Findings:**
|
||||||
|
- Clear summary of analysis and recommendations
|
||||||
|
- Expected benefits and potential risks
|
||||||
|
- Implementation plan and timeline
|
||||||
|
|
||||||
|
**Seek Approval:**
|
||||||
|
- Explicit user confirmation for significant methodology changes
|
||||||
|
- Discussion of concerns or alternative approaches
|
||||||
|
- Agreement on implementation priorities
|
||||||
|
|
||||||
|
### 7. Implementation and Tracking
|
||||||
|
Once approved:
|
||||||
|
|
||||||
|
**Apply Changes:**
|
||||||
|
- Update relevant persona files, tasks, and templates
|
||||||
|
- Modify workflow documentation
|
||||||
|
- Update CLAUDE.md if necessary
|
||||||
|
|
||||||
|
**Document Changes:**
|
||||||
|
- Record all changes in the improvement log
|
||||||
|
- Note expected outcomes and success metrics
|
||||||
|
- Set up tracking for effectiveness validation
|
||||||
|
|
||||||
|
**Validate Improvements:**
|
||||||
|
- Monitor effectiveness of implemented changes
|
||||||
|
- Collect feedback on new approaches
|
||||||
|
- Plan follow-up retrospectives to assess impact
|
||||||
|
|
||||||
|
## Output Format
|
||||||
|
|
||||||
|
Create a structured retrospective report with:
|
||||||
|
|
||||||
|
1. **Executive Summary** - Key findings and recommendations
|
||||||
|
2. **Performance Metrics** - Quantified analysis of recent phase
|
||||||
|
3. **Pattern Analysis** - Successful and problematic patterns identified
|
||||||
|
4. **Improvement Proposals** - Specific, prioritized recommendations
|
||||||
|
5. **Implementation Plan** - Steps to apply approved changes
|
||||||
|
6. **Success Metrics** - How to measure improvement effectiveness
|
||||||
|
|
||||||
|
## Integration with BMAD Evolution
|
||||||
|
|
||||||
|
This task directly supports the self-improving nature of the BMAD framework by:
|
||||||
|
- Providing systematic methodology improvement
|
||||||
|
- Ensuring continuous optimization based on real-world usage
|
||||||
|
- Maintaining quality while enabling rapid evolution
|
||||||
|
- Creating a feedback loop for all BMAD components
|
||||||
|
|
||||||
|
Execute this task regularly to ensure the BMAD methodology continues to evolve and improve with each project experience.
|
||||||
|
|
@ -0,0 +1,197 @@
|
||||||
|
# Persona Optimization Task
|
||||||
|
|
||||||
|
## Purpose
|
||||||
|
Systematically analyze and optimize individual BMAD persona performance to enhance their effectiveness and capabilities.
|
||||||
|
|
||||||
|
## When to Execute
|
||||||
|
- After completing 3-5 projects with a specific persona
|
||||||
|
- When persona performance metrics indicate suboptimal results
|
||||||
|
- Following user feedback about persona effectiveness
|
||||||
|
- Before major methodology updates or overhauls
|
||||||
|
|
||||||
|
## Optimization Process
|
||||||
|
|
||||||
|
### 1. Persona Performance Analysis
|
||||||
|
|
||||||
|
**Current Capability Assessment:**
|
||||||
|
- Review recent deliverable quality and consistency
|
||||||
|
- Analyze time-to-completion metrics for typical tasks
|
||||||
|
- Evaluate user satisfaction ratings with persona interactions
|
||||||
|
- Assess alignment between persona outputs and requirements
|
||||||
|
|
||||||
|
**Strength Identification:**
|
||||||
|
- Which persona capabilities consistently produce excellent results?
|
||||||
|
- What unique value does this persona provide to the BMAD workflow?
|
||||||
|
- Which techniques or approaches work exceptionally well?
|
||||||
|
|
||||||
|
**Weakness Identification:**
|
||||||
|
- Where does the persona consistently struggle or underperform?
|
||||||
|
- What types of tasks or requirements create difficulties?
|
||||||
|
- Which outputs require frequent iteration or clarification?
|
||||||
|
|
||||||
|
### 2. Instruction Effectiveness Review
|
||||||
|
|
||||||
|
**Core Principles Analysis:**
|
||||||
|
- Are the existing principles still relevant and effective?
|
||||||
|
- Do any principles conflict or create confusion?
|
||||||
|
- Are there missing principles that would improve performance?
|
||||||
|
|
||||||
|
**Operational Instructions Review:**
|
||||||
|
- Which instructions are followed consistently vs. ignored?
|
||||||
|
- Are there unclear or ambiguous instructions?
|
||||||
|
- Do instructions scale well across different project types?
|
||||||
|
|
||||||
|
**Example and Template Quality:**
|
||||||
|
- Are provided examples still current and helpful?
|
||||||
|
- Do templates facilitate or hinder optimal performance?
|
||||||
|
- Are there gaps in guidance for common scenarios?
|
||||||
|
|
||||||
|
### 3. Capability Gap Analysis
|
||||||
|
|
||||||
|
**Missing Skills or Knowledge:**
|
||||||
|
- What capabilities would significantly improve persona effectiveness?
|
||||||
|
- Are there emerging patterns or technologies the persona should understand?
|
||||||
|
- What domain expertise updates are needed?
|
||||||
|
|
||||||
|
**Process Improvement Opportunities:**
|
||||||
|
- How could the persona's workflow be optimized?
|
||||||
|
- What additional tools or frameworks would be helpful?
|
||||||
|
- Which handoff processes could be improved?
|
||||||
|
|
||||||
|
**Quality Enhancement Potential:**
|
||||||
|
- What would elevate good outputs to exceptional ones?
|
||||||
|
- How could consistency be improved across different contexts?
|
||||||
|
- What would reduce the need for iterations or refinements?
|
||||||
|
|
||||||
|
### 4. Optimization Strategy Development
|
||||||
|
|
||||||
|
**Priority Categorization:**
|
||||||
|
- Critical: Issues that significantly impact persona effectiveness
|
||||||
|
- Important: Improvements that would provide substantial benefits
|
||||||
|
- Nice-to-have: Enhancements for incremental optimization
|
||||||
|
|
||||||
|
**Implementation Complexity:**
|
||||||
|
- Simple: Changes to existing instructions or principles
|
||||||
|
- Moderate: Addition of new capabilities or frameworks
|
||||||
|
- Complex: Fundamental restructuring of persona approach
|
||||||
|
|
||||||
|
**Risk Assessment:**
|
||||||
|
- What are the potential negative impacts of proposed changes?
|
||||||
|
- How can changes be tested before full implementation?
|
||||||
|
- What rollback procedures are needed if changes don't work?
|
||||||
|
|
||||||
|
### 5. Optimization Implementation
|
||||||
|
|
||||||
|
**Instruction Updates:**
|
||||||
|
- Refine existing principles for clarity and effectiveness
|
||||||
|
- Add new principles to address identified gaps
|
||||||
|
- Remove or modify counterproductive instructions
|
||||||
|
|
||||||
|
**Capability Enhancements:**
|
||||||
|
- Integrate new domain knowledge or techniques
|
||||||
|
- Add frameworks or tools to improve performance
|
||||||
|
- Enhance templates and examples for better guidance
|
||||||
|
|
||||||
|
**Process Improvements:**
|
||||||
|
- Streamline workflows for better efficiency
|
||||||
|
- Improve handoff procedures with other personas
|
||||||
|
- Add quality control checkpoints where needed
|
||||||
|
|
||||||
|
### 6. Testing and Validation
|
||||||
|
|
||||||
|
**Controlled Testing:**
|
||||||
|
- Test optimized persona on representative tasks
|
||||||
|
- Compare performance against baseline metrics
|
||||||
|
- Gather feedback from users and stakeholders
|
||||||
|
|
||||||
|
**A/B Comparison:**
|
||||||
|
- Run parallel tests with old vs. new persona versions
|
||||||
|
- Measure differences in key performance indicators
|
||||||
|
- Validate that optimizations deliver expected benefits
|
||||||
|
|
||||||
|
**Iterative Refinement:**
|
||||||
|
- Make adjustments based on testing feedback
|
||||||
|
- Fine-tune optimizations for better results
|
||||||
|
- Ensure changes integrate well with overall BMAD workflow
|
||||||
|
|
||||||
|
## Specific Persona Optimization Areas
|
||||||
|
|
||||||
|
### Analyst Optimization Focus
|
||||||
|
- Research methodology effectiveness
|
||||||
|
- Brainstorming facilitation techniques
|
||||||
|
- Project brief clarity and completeness
|
||||||
|
- Pattern recognition for successful ideation
|
||||||
|
|
||||||
|
### PM Optimization Focus
|
||||||
|
- Requirements gathering efficiency
|
||||||
|
- PRD structure and clarity
|
||||||
|
- Epic and story definition quality
|
||||||
|
- Stakeholder alignment effectiveness
|
||||||
|
|
||||||
|
### Architect Optimization Focus
|
||||||
|
- Technical decision-making speed and accuracy
|
||||||
|
- Architecture documentation clarity
|
||||||
|
- Technology selection rationale
|
||||||
|
- Implementation guidance quality
|
||||||
|
|
||||||
|
### Design Architect Optimization Focus
|
||||||
|
- UI/UX specification completeness
|
||||||
|
- Frontend architecture clarity
|
||||||
|
- Design-to-development handoff effectiveness
|
||||||
|
- User experience validation techniques
|
||||||
|
|
||||||
|
### PO Optimization Focus
|
||||||
|
- Cross-artifact validation effectiveness
|
||||||
|
- Story prioritization accuracy
|
||||||
|
- Change management capability
|
||||||
|
- Quality assurance thoroughness
|
||||||
|
|
||||||
|
### Development Persona Optimization Focus
|
||||||
|
- Code quality and consistency
|
||||||
|
- Implementation speed and accuracy
|
||||||
|
- Technical debt minimization
|
||||||
|
- Documentation and testing practices
|
||||||
|
|
||||||
|
## Optimization Output
|
||||||
|
|
||||||
|
### Updated Persona Documentation
|
||||||
|
- Revised persona file with optimized instructions
|
||||||
|
- Updated core principles and capabilities
|
||||||
|
- Enhanced examples and templates
|
||||||
|
- Improved workflow and process guidance
|
||||||
|
|
||||||
|
### Performance Improvement Plan
|
||||||
|
- Specific metrics to track improvement
|
||||||
|
- Timeline for optimization validation
|
||||||
|
- Success criteria for optimization effectiveness
|
||||||
|
- Rollback plan if optimizations prove problematic
|
||||||
|
|
||||||
|
### Integration Updates
|
||||||
|
- Modifications to how persona interacts with others
|
||||||
|
- Updates to handoff procedures and expectations
|
||||||
|
- Changes to overall BMAD workflow if needed
|
||||||
|
- Communication plan for optimization rollout
|
||||||
|
|
||||||
|
## Success Metrics
|
||||||
|
|
||||||
|
**Quantitative Improvements:**
|
||||||
|
- Reduced time-to-completion for typical tasks
|
||||||
|
- Increased user satisfaction ratings
|
||||||
|
- Improved deliverable quality scores
|
||||||
|
- Decreased iteration cycles needed
|
||||||
|
|
||||||
|
**Qualitative Enhancements:**
|
||||||
|
- More consistent high-quality outputs
|
||||||
|
- Better alignment with user expectations
|
||||||
|
- Smoother integration with BMAD workflow
|
||||||
|
- Enhanced capability in challenging scenarios
|
||||||
|
|
||||||
|
## Continuous Optimization
|
||||||
|
|
||||||
|
This task should be executed regularly as part of the BMAD framework's self-improvement process. Each optimization cycle should:
|
||||||
|
- Build on previous improvements
|
||||||
|
- Adapt to changing project requirements
|
||||||
|
- Incorporate new best practices and techniques
|
||||||
|
- Maintain the persona's unique value proposition
|
||||||
|
|
||||||
|
The goal is continuous evolution toward increasingly effective and valuable BMAD personas that deliver exceptional results across diverse project contexts.
|
||||||
|
|
@ -6,7 +6,7 @@ This document tracks all improvements, changes, and evolution of the BMAD method
|
||||||
|
|
||||||
### v1.0 - Initial Self-Improving Framework (Milestone 1)
|
### v1.0 - Initial Self-Improving Framework (Milestone 1)
|
||||||
**Date**: Initial Implementation
|
**Date**: Initial Implementation
|
||||||
**Commit**: TBD
|
**Commit**: a6f1bf7 - "Milestone 1: Initialize Self-Improving BMAD Framework"
|
||||||
|
|
||||||
#### Changes Made:
|
#### Changes Made:
|
||||||
- Transformed static BMAD framework into self-improving system
|
- Transformed static BMAD framework into self-improving system
|
||||||
|
|
@ -24,6 +24,33 @@ This document tracks all improvements, changes, and evolution of the BMAD method
|
||||||
- Baseline established for future comparison
|
- Baseline established for future comparison
|
||||||
- Framework prepared for adaptive learning
|
- Framework prepared for adaptive learning
|
||||||
|
|
||||||
|
### v2.0 - Meta-Improvement Infrastructure (Milestone 2)
|
||||||
|
**Date**: Phase 2 Implementation
|
||||||
|
**Commit**: TBD
|
||||||
|
|
||||||
|
#### Changes Made:
|
||||||
|
- Enhanced personas with self-improvement principles and capabilities
|
||||||
|
- Created comprehensive improvement tracking and measurement systems
|
||||||
|
- Added methodology optimization tasks for systematic enhancement
|
||||||
|
- Implemented inter-persona feedback loops for collaborative learning
|
||||||
|
|
||||||
|
#### Key Improvements:
|
||||||
|
- **Self-Improving Personas**: All personas now have built-in learning and optimization capabilities
|
||||||
|
- **Systematic Measurement**: Comprehensive effectiveness tracking with velocity, quality, and satisfaction metrics
|
||||||
|
- **Optimization Tasks**: Structured approaches for persona improvement and methodology enhancement
|
||||||
|
- **Collaborative Learning**: Feedback loops between personas enable continuous workflow optimization
|
||||||
|
|
||||||
|
#### New Capabilities Added:
|
||||||
|
- Methodology Retrospective Task - systematic analysis of completed phases
|
||||||
|
- Effectiveness Measurement Task - comprehensive metrics tracking system
|
||||||
|
- Persona Optimization Task - individual persona enhancement framework
|
||||||
|
- Inter-Persona Feedback Task - collaborative improvement between personas
|
||||||
|
|
||||||
|
#### Impact Metrics:
|
||||||
|
- Infrastructure ready for automated improvement detection
|
||||||
|
- Personas equipped with self-optimization capabilities
|
||||||
|
- Measurement systems in place for data-driven enhancement
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
## Improvement Templates
|
## Improvement Templates
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue