9.9 KiB
| description | argument-hint | allowed-tools | ||||||
|---|---|---|---|---|---|---|---|---|
| Create comprehensive test plans for any functionality (epics, stories, features, custom) | [epic-3] [story-2.1] [feature-login] [custom-functionality] [--overwrite] |
|
⚠️ GENERAL-PURPOSE COMMAND - Works with any project
Documentation directories are detected dynamically (docs/, documentation/, wiki/)
Output directory is detected dynamically (workspace/testing/plans, test-plans, .)
Override with CREATE_TEST_PLAN_OUTPUT_DIR environment variable if needed
📋 Test Plan Creator - High Context Analysis
Argument Processing
Target functionality: "$ARGUMENTS"
Parse functionality identifier:
const arguments = "$ARGUMENTS";
const functionalityPattern = /(?:epic-[\d]+(?:\.[\d]+)?|story-[\d]+(?:\.[\d]+)?|feature-[\w-]+|[\w-]+)/g;
const functionalityMatch = arguments.match(functionalityPattern)?.[0] || "custom-functionality";
const overwrite = arguments.includes("--overwrite");
Target: ${functionalityMatch}
Overwrite existing: ${overwrite ? "Yes" : "No"}
Test Plan Creation Process
Step 0: Detect Project Structure
# ============================================
# DYNAMIC DIRECTORY DETECTION (Project-Agnostic)
# ============================================
# Detect documentation directories
DOCS_DIRS=""
for dir in "docs" "documentation" "wiki" "spec" "specifications"; do
if [[ -d "$dir" ]]; then
DOCS_DIRS="$DOCS_DIRS $dir"
fi
done
if [[ -z "$DOCS_DIRS" ]]; then
echo "⚠️ No documentation directory found (docs/, documentation/, etc.)"
echo " Will search current directory for documentation files"
DOCS_DIRS="."
fi
echo "📁 Documentation directories: $DOCS_DIRS"
# Detect output directory (allow env override)
if [[ -n "$CREATE_TEST_PLAN_OUTPUT_DIR" ]]; then
PLANS_DIR="$CREATE_TEST_PLAN_OUTPUT_DIR"
echo "📁 Using override output dir: $PLANS_DIR"
else
PLANS_DIR=""
for dir in "workspace/testing/plans" "test-plans" "testing/plans" "tests/plans"; do
if [[ -d "$dir" ]]; then
PLANS_DIR="$dir"
break
fi
done
# Create in first available parent
if [[ -z "$PLANS_DIR" ]]; then
for dir in "workspace/testing/plans" "test-plans" "testing/plans"; do
PARENT_DIR=$(dirname "$dir")
if [[ -d "$PARENT_DIR" ]] || mkdir -p "$PARENT_DIR" 2>/dev/null; then
mkdir -p "$dir" 2>/dev/null && PLANS_DIR="$dir" && break
fi
done
# Ultimate fallback
if [[ -z "$PLANS_DIR" ]]; then
PLANS_DIR="./test-plans"
mkdir -p "$PLANS_DIR"
fi
fi
echo "📁 Test plans directory: $PLANS_DIR"
fi
Step 1: Check for Existing Plan
Check if test plan already exists:
planFile="$PLANS_DIR/${functionalityMatch}-test-plan.md"
if [[ -f "$planFile" && "$overwrite" != true ]]; then
echo "⚠️ Test plan already exists: $planFile"
echo "Use --overwrite to replace existing plan"
exit 1
fi
Step 2: Comprehensive Requirements Analysis
FULL CONTEXT ANALYSIS - This is where the high-context work happens:
Document Discovery: Use Grep and Read tools to find ALL relevant documentation:
- Search
docs/prd/*${functionalityMatch}*.md - Search
docs/stories/*${functionalityMatch}*.md - Search
docs/features/*${functionalityMatch}*.md - Search project files for functionality references
- Analyze any custom specifications provided
Requirements Extraction: For EACH discovered document, extract:
- Acceptance Criteria: All AC patterns (AC X.X.X, Given-When-Then, etc.)
- User Stories: "As a...I want...So that..." patterns
- Integration Points: System interfaces, APIs, dependencies
- Success Metrics: Performance thresholds, quality requirements
- Risk Areas: Edge cases, potential failure modes
- Business Logic: Domain-specific requirements (like Mike Israetel methodology)
Context Integration:
- Cross-reference requirements across multiple documents
- Identify dependencies between different acceptance criteria
- Map user workflows that span multiple components
- Understand system architecture context
Step 3: Test Scenario Design
Mode-Specific Scenario Planning: For each testing mode (automated/interactive/hybrid), design:
Automated Scenarios:
- Browser automation sequences using MCP tools
- API endpoint validation workflows
- Performance measurement checkpoints
- Error condition testing scenarios
Interactive Scenarios:
- Human-guided test procedures
- User experience validation flows
- Qualitative assessment activities
- Accessibility and usability evaluation
Hybrid Scenarios:
- Automated setup + manual validation
- Quantitative collection + qualitative interpretation
- Parallel automated/manual execution paths
Step 4: Validation Criteria Definition
Measurable Success Criteria: For each scenario, define:
- Functional Validation: Feature behavior correctness
- Performance Validation: Response times, resource usage
- Quality Validation: User experience, accessibility, reliability
- Integration Validation: Cross-system communication, data flow
Evidence Requirements:
- Automated Evidence: Screenshots, logs, metrics, API responses
- Manual Evidence: User feedback, qualitative observations
- Hybrid Evidence: Combined data + human interpretation
Step 5: Agent Prompt Generation
Specialized Agent Instructions: Create detailed prompts for each subagent that include:
- Specific context from the requirements analysis
- Detailed instructions for their specialized role
- Expected input/output formats
- Integration points with other agents
Step 6: Test Plan File Generation
Create comprehensive test plan file:
# Test Plan: ${functionalityMatch}
**Created**: $(date)
**Target**: ${functionalityMatch}
**Context**: [Summary of analyzed documentation]
## Requirements Analysis
### Source Documents
- [List of all documents analyzed]
- [Cross-references and dependencies identified]
### Acceptance Criteria
[All extracted ACs with full context]
### User Stories
[All user stories requiring validation]
### Integration Points
[System interfaces and dependencies]
### Success Metrics
[Performance thresholds and quality requirements]
### Risk Areas
[Edge cases and potential failure modes]
## Test Scenarios
### Automated Test Scenarios
[Detailed browser automation and API test scenarios]
### Interactive Test Scenarios
[Human-guided testing procedures and UX validation]
### Hybrid Test Scenarios
[Combined automated + manual approaches]
## Validation Criteria
### Success Thresholds
[Measurable pass/fail criteria for each scenario]
### Evidence Requirements
[What evidence proves success or failure]
### Quality Gates
[Performance, usability, and reliability standards]
## Agent Execution Prompts
### Requirements Analyzer Prompt
Context: ${functionalityMatch} testing based on comprehensive requirements analysis Task: [Specific instructions based on discovered documentation] Expected Output: [Structured requirements summary]
### Scenario Designer Prompt
Context: Transform ${functionalityMatch} requirements into executable test scenarios Task: [Mode-specific scenario generation instructions] Expected Output: [Test scenario definitions]
### Validation Planner Prompt
Context: Define success criteria for ${functionalityMatch} validation
Task: [Validation criteria and evidence requirements]
Expected Output: [Comprehensive validation plan]
### Browser Executor Prompt
Context: Execute automated tests for ${functionalityMatch} Task: [Browser automation and performance testing] Expected Output: [Execution results and evidence]
### Interactive Guide Prompt
Context: Guide human testing of ${functionalityMatch} Task: [User experience and qualitative validation] Expected Output: [Interactive session results]
### Evidence Collector Prompt
Context: Aggregate all ${functionalityMatch} testing evidence Task: [Evidence compilation and organization] Expected Output: [Comprehensive evidence package]
### BMAD Reporter Prompt
Context: Generate final report for ${functionalityMatch} testing Task: [Analysis and actionable recommendations] Expected Output: [BMAD-format final report]
## Execution Notes
### Testing Modes
- **Automated**: Focus on browser automation, API validation, performance
- **Interactive**: Emphasize user experience, usability, qualitative insights
- **Hybrid**: Combine automated metrics with human interpretation
### Context Preservation
- All agents receive full context from this comprehensive analysis
- Cross-references maintained between requirements and scenarios
- Integration dependencies clearly mapped
### Reusability
- Plan can be executed multiple times with different modes
- Scenarios can be updated independently
- Agent prompts can be refined based on results
---
*Test Plan Created: $(date)*
*High-Context Analysis: Complete requirements discovery and scenario design*
*Ready for execution via /user_testing ${functionalityMatch}*
Completion
Display results:
✅ Test Plan Created Successfully!
================================================================
📋 Plan: ${functionalityMatch}-test-plan.md
📁 Location: $PLANS_DIR/
🎯 Target: ${functionalityMatch}
📊 Analysis: Complete requirements and scenario design
================================================================
🚀 Next Steps:
1. Review the comprehensive test plan in $PLANS_DIR/
2. Execute tests using: /user_testing ${functionalityMatch} --mode=[automated|interactive|hybrid]
3. Test plan can be reused and refined for multiple execution sessions
4. Plan includes specialized prompts for all 7 subagents
📝 Plan Contents:
- Complete requirements analysis with full context
- Mode-specific test scenarios (automated/interactive/hybrid)
- Measurable validation criteria and evidence requirements
- Specialized agent prompts with comprehensive context
- Execution guidance and quality gates
Test Plan Creator v1.0 - High Context Analysis for Comprehensive Testing