* feat(tasks): convert review-adversarial-general from XML task to native skill Convert the simplest core task (review-adversarial-general.xml) from type:task XML format to type:skill markdown format. This establishes the pattern for converting remaining XML tasks to self-contained skills. - Convert XML task to workflow.md with frontmatter, role, execution steps - Add type:skill manifest for verbatim directory copying - Extend manifest-generator getTasksFromDir to recurse into subdirectories and detect type:skill entries (mirrors existing workflow skill detection) - Update cross-references in quick-dev-new-preview, quick-dev, quick-spec - Update module-help.csv to use skill: prefix * refactor: replace file path references with skill name invocations Consumers of review-adversarial-general now invoke by skill name instead of loading via _bmad/ file path. Removes the indirection variable from frontmatter and inlines the skill name directly. * refactor(installer): scan tasks/ for type:skill entries Teach collectWorkflows to also scan the tasks/ subdirectory for type:skill entries. Skills can live anywhere in the source tree — the workflow scanner just needs to look in more places. * fix: update stale task terminology to skill after format conversion Address review findings from PR #1857: replace remaining "task" references with "skill" in workflow steps and test documentation. Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com> --------- Co-authored-by: Claude Opus 4.6 <noreply@anthropic.com> |
||
|---|---|---|
| .. | ||
| README.md | ||
| sample-content.md | ||
| test-cases.yaml | ||
README.md
Adversarial Review Test Suite
Tests for the also_consider optional input in the bmad-review-adversarial-general skill.
Purpose
Evaluate whether the also_consider input gently nudges the reviewer toward specific areas without overriding normal adversarial analysis.
Test Content
All tests use sample-content.md - a deliberately imperfect User Authentication API doc with:
- Vague error handling section
- Missing rate limit details
- No token expiration info
- Password in plain text example
- Missing authentication headers
- No error response examples
Running Tests
For each test case in test-cases.yaml, invoke the adversarial review skill.
Manual Test Invocation
Review this content using the adversarial review skill:
<content>
[paste sample-content.md]
</content>
<also_consider>
[paste items from test case, or omit for TC01]
</also_consider>
Evaluation Criteria
For each test, note:
- Total findings - Still hitting ~10 issues?
- Distribution - Are findings spread across concerns or clustered?
- Relevance - Do findings relate to
also_consideritems when provided? - Balance - Are
also_considerfindings elevated over others, or naturally mixed? - Quality - Are findings actionable regardless of source?
Expected Outcomes
- TC01 (baseline): Generic spread of findings
- TC02-TC05 (domain-focused): Some findings align with domain, others still organic
- TC06 (single item): Light influence, not dominant
- TC07 (vague items): Minimal change from baseline
- TC08 (specific items): Direct answers if gaps exist
- TC09 (mixed): Balanced across domains
- TC10 (contradictory): Graceful handling