BMAD-METHOD/expansion-packs/bmad-technical-writing/tasks/incorporate-reviewer-feedba...

9.1 KiB

Incorporate Reviewer Feedback


task: id: incorporate-reviewer-feedback name: Systematically Incorporate Reviewer Feedback description: Process and address technical reviewer, publisher, and beta reader feedback systematically persona_default: book-analyst inputs: - reviewer_feedback (technical review comments, publisher requests, beta reader notes) - affected_chapters steps: - Collect all reviewer feedback from all sources (technical, publisher, beta readers) - Categorize feedback by severity (critical/must-fix, important/should-fix, optional/nice-to-have) - Create feedback tracking log with status for each item - Address critical issues first (technical errors, broken code, security issues) - Fix important issues (clarity problems, missing examples, structural issues) - Consider optional suggestions (enhancements, additional topics, style preferences) - Test all code changes from feedback - Update text for clarity improvements requested - Track completion status in feedback log - Generate feedback-resolution-log documenting all changes - Run execute-checklist.md with existing-book-integration-checklist.md output: docs/feedback/{{book_title}}-feedback-resolution-log.md


Purpose

This task provides a systematic approach to processing reviewer feedback from technical reviewers, publishers, and beta readers. Ensures all feedback is triaged, addressed appropriately, and tracked to completion.

Prerequisites

Before starting this task:

  • Reviewer feedback collected from all sources
  • Chapters are in reviewable state
  • Testing environment set up for code changes
  • Understanding of feedback priorities (which issues are critical)

Workflow Steps

1. Collect All Reviewer Feedback

Gather feedback from all sources:

Technical Reviewer Feedback:

  • Technical accuracy issues
  • Code errors or improvements
  • Misleading explanations
  • Missing prerequisites
  • Incorrect terminology

Publisher Feedback:

  • Format compliance issues
  • Style guide violations
  • Length adjustments needed
  • Market positioning changes
  • Legal/licensing concerns

Beta Reader Feedback:

  • Clarity problems
  • Confusing sections
  • Missing examples
  • Difficulty level issues
  • Typos and errors

Consolidate into a single master feedback list.

2. Categorize Feedback by Severity

Triage each feedback item into priority categories:

Critical (Must-Fix):

  • Technical errors (incorrect information)
  • Broken code examples (won't run)
  • Security vulnerabilities
  • Legal/licensing issues
  • Publisher blocking issues (won't publish without fix)
  • Major clarity problems (readers can't follow)

Important (Should-Fix):

  • Unclear explanations (could be clearer)
  • Missing examples (would help understanding)
  • Structural issues (better organization possible)
  • Incomplete coverage (topic needs expansion)
  • Style inconsistencies
  • Minor technical inaccuracies

Nice-to-Have (Optional):

  • Style preferences (subjective improvements)
  • Additional topics (scope expansion)
  • Enhancement suggestions
  • Alternative explanations
  • Personal preferences

3. Create Feedback Tracking Log

Build a structured tracking system:

ID Chapter Severity Issue Requested By Status Resolution Date
F001 Ch 3 Critical Code won't run Python 3.12 Tech Review Done Fixed import 2024-01-15
F002 Ch 5 Important Unclear JWT explanation Beta Reader Done Added example 2024-01-16
F003 Ch 7 Optional Add async/await example Tech Review Deferred Future edition 2024-01-16

This provides visibility into progress and ensures nothing is missed.

4. Address Critical Issues First

Start with must-fix items:

For Technical Errors:

  • Verify the error (confirm it's incorrect)
  • Research the correct information
  • Update text and code
  • Test updated code
  • Add verification note to tracking log

For Broken Code:

  • Reproduce the issue
  • Fix the code
  • Test on target version(s)
  • Verify output is correct
  • Update text if output changed

For Security Issues:

  • Assess severity (CVSS score if applicable)
  • Fix immediately
  • Add security note if appropriate
  • Test fix thoroughly
  • Document in change log

For Publisher Blocking Issues:

  • Understand exact requirement
  • Implement change
  • Verify compliance
  • Get publisher confirmation
  • Mark resolved

Do not proceed to lower-priority items until all critical issues are resolved.

5. Fix Important Issues

Address should-fix items systematically:

For Clarity Problems:

  • Identify specific unclear section
  • Rewrite for clarity
  • Add examples if needed
  • Get second opinion (beta reader, colleague)
  • Update tracking log

For Missing Examples:

  • Understand what example is needed
  • Design example that teaches the concept
  • Write and test code
  • Integrate into chapter
  • Verify it improves understanding

For Structural Issues:

  • Assess reorganization impact
  • Plan structural change
  • Reorganize content
  • Update cross-references
  • Verify learning flow still works

For Incomplete Coverage:

  • Determine scope of addition
  • Write additional content
  • Test any new code
  • Integrate smoothly
  • Ensure doesn't bloat chapter excessively

6. Consider Optional Suggestions

Evaluate nice-to-have items carefully:

Decision Criteria:

  • Does it improve reader experience?
  • Is it within scope of current edition?
  • Do I have time/space for this?
  • Does it align with book goals?

Actions:

  • Implement: If valuable and feasible
  • Defer: If good idea but not for this edition (document for next edition)
  • Decline: If not aligned with book goals (document reason)

Document all decisions in tracking log, even for declined items.

7. Test All Code Changes

For every code change made from feedback:

  • Test code runs successfully
  • Test on target version(s)
  • Verify output matches text
  • Check for new errors or warnings
  • Run regression tests (ensure other examples still work)
  • Update code repository

No code changes should be marked complete without testing.

8. Update Text for Clarity

For text improvements from feedback:

  • Rewrite unclear sections
  • Add clarifying examples
  • Improve explanations
  • Fix terminology inconsistencies
  • Verify technical accuracy
  • Ensure voice/tone consistency

Use extracted code patterns and style guide to maintain consistency.

9. Track Completion Status

Update feedback tracking log continuously:

  • Mark items as "In Progress" when starting
  • Mark as "Done" when complete and tested
  • Mark as "Deferred" if postponing to next edition
  • Mark as "Declined" if not implementing (with reason)
  • Add completion date
  • Add resolution notes

This creates accountability and progress visibility.

10. Generate Feedback Resolution Log

Create comprehensive document summarizing all feedback processing:

# Feedback Resolution Log - [Book Title]

## Summary
- Total feedback items: 47
- Critical (resolved): 8/8
- Important (resolved): 23/25 (2 deferred)
- Optional (resolved): 7/14 (4 deferred, 3 declined)

## Critical Issues Resolved
[List with details]

## Important Issues Resolved
[List with details]

## Deferred Items
[List with rationale and target edition]

## Declined Items
[List with rationale]

## Code Changes
[List all code changes made]

## Text Changes
[List major text revisions]

## Reviewer Acknowledgments
[Thank reviewers]

This document provides transparency and completeness.

11. Run Integration Checklist

Use execute-checklist.md with existing-book-integration-checklist.md to ensure:

  • Changes maintain consistency with existing content
  • Voice and tone are consistent
  • Code patterns are followed
  • Cross-references are accurate
  • Learning flow is maintained

Success Criteria

A completed feedback incorporation should have:

  • All feedback collected from all sources
  • Feedback categorized by severity
  • Tracking log created and maintained
  • All critical issues resolved
  • All important issues addressed or consciously deferred
  • Optional items evaluated (implement, defer, or decline)
  • All code changes tested
  • Text clarity improvements made
  • Completion status tracked for every item
  • Feedback resolution log generated
  • Integration checklist passed
  • No blocking issues remain

Common Pitfalls to Avoid

  • Ignoring low-severity feedback: Track and evaluate all feedback, even if declining
  • No prioritization: Must address critical items first
  • Scope creep: Optional items can expand scope significantly - be disciplined
  • Poor tracking: Without tracking, items get missed
  • Untested changes: All code changes must be tested
  • Inconsistent voice: Text changes must match existing style
  • No documentation: Document what changed and why

Next Steps

After incorporating feedback:

  1. Send resolution log to reviewers for confirmation
  2. Request final approval from technical reviewer
  3. Get publisher sign-off on critical fixes
  4. Proceed to final editorial review
  5. Prepare for publication
  6. Archive deferred items for next edition planning