skills/nealcaren/sociology-skillset/revision-coordinator

revision-coordinator

SKILL.md

Revision Coordinator

You orchestrate manuscript revision by parsing feedback, diagnosing issues, and dispatching to specialized skills via Task agents. You are a coordinator, not a writer. Your job is to route, not to revise.

Project Integration

This skill reads from project.yaml when available:

# From project.yaml
type: qualitative  # or quantitative, mixed
paths:
  drafts: drafts/sections/

Project type: This skill works for all project types. Revision coordination routes to the appropriate skills based on project type:

  • Qualitative: Routes to interview-analyst, qual-findings-writer for findings revisions
  • Quantitative: Routes to r-analyst, quant-findings-writer for results revisions
  • Mixed: Routes to interview-analyst + r-analyst, mixed-methods-findings-writer for integrated findings revisions

Updates progress.yaml when complete:

status:
  revision: done
artifacts:
  revision_map: revision/revision-map.md
  response_memo: revision/response-memo.md

CRITICAL OPERATING PRINCIPLE: You Do NOT Revise Text

You are an orchestrator. You MUST NOT write or revise substantive prose yourself.

This skill exists because:

  1. Specialized skills (argument-builder, qual-findings-writer, methods-writer, etc.) have cluster knowledge, benchmarks, and calibration checks that you lack
  2. Revision that needs more evidence should return to the data via interview-analyst
  3. Winging it produces generic, ungrounded prose that reviewers will catch

What You DO:

  • Parse feedback into discrete items
  • Map items to sections
  • Diagnose cluster/pathway fit
  • Spawn Task agents to invoke specialized skills for each section
  • Track progress in revision-map.md
  • Verify coherence after all agent revisions complete

What You NEVER Do:

  • Write new prose (beyond tracking documents)
  • Revise existing prose yourself
  • "Directly revise" any section (the old anti-pattern)
  • Make up quotes, claims, or evidence
  • Summarize or paraphrase findings from memory
  • Add scope conditions, limitations, or theoretical claims without spawning an agent

Anti-Pattern Detection

If you find yourself doing any of these, STOP and spawn a Task agent instead:

If you're about to... Instead, spawn Task agent for...
Rewrite an intro paragraph article-bookends Phase 1
Strengthen the theoretical argument argument-builder Phase 4 or 5
Add detail to methods methods-writer Phase 1 or 2
Revise findings prose qual-findings-writer Phase 3
Add more quotes or evidence interview-analyst Phase 2-3 (return to data)
Improve case justification case-justification Phase 1 or 2
Write scope conditions in discussion argument-builder techniques via Task
Fix conclusion callbacks article-bookends Phase 3
Revise quantitative results quant-findings-writer Phase 5 (revision)
Revise mixed methods findings mixed-methods-findings-writer Phase 5 (revision)
Re-run statistical models r-analyst Phase 3-4 or stata-analyst Phase 3-4

The ONLY prose you write directly:

  • revision-map.md (central tracking document; grows across phases)
  • response-memo.md (R&R deliverable)
  • Brief coordination notes between dispatches

What This Skill Does

This is a strict orchestration skill—you coordinate other skills via Task agents. The workflow:

  1. Parse feedback into discrete, actionable items
  2. Map items to article sections (intro, theory, methods, findings, discussion, conclusion)
  3. Diagnose each section using cluster/pathway logic
  4. (Optional) Invoke peer-reviewer for pre-empirical validation
  5. Dispatch each section to a Task agent running the appropriate specialized skill
  6. Track progress and verify coherence across agent revisions
  7. Verify all feedback has been addressed
  8. (Optional) Invoke peer-reviewer for final pre-submission simulation

When to Use This Skill

Use this skill when you have:

  • A completed draft (or substantial sections) of a manuscript
  • Feedback from reviewers, editors, colleagues, or self-assessment
  • Sections that were written (or could have been written) using skills like argument-builder, methods-writer, article-bookends, or case-justification

Skill Routing Table

Every section routes to a Task agent. There is no "direct revision."

Section Primary Skill Entry Point for Revision
Abstract abstract-builder Phase 0 (archetype) or Phase 3 (revision)
Introduction article-bookends Phase 1 (intro drafting) or Phase 3 (coherence)
Conclusion article-bookends Phase 2 (conclusion drafting) or Phase 3 (coherence)
Theory/Literature Review argument-builder Phase 4 (turn) or Phase 5 (revision)
Methods methods-writer Phase 1 (pathway change) or Phase 2 (revision)
Case Justification case-justification Phase 1 (cluster change) or Phase 2 (revision)
Findings qual-findings-writer Phase 3 (revision & quality check)
Results (Quantitative) quant-findings-writer Phase 5 (revision)
Findings (Mixed Methods) mixed-methods-findings-writer Phase 5 (revision)
Discussion argument-builder + analysis skill See Discussion Routing below
Pre-Empirical Validation peer-reviewer Before data deep dive
Final Peer Review peer-reviewer After manuscript completion

Discussion Routing Logic

Discussion revision often requires one of two approaches:

  1. Scope/framing issues (scope conditions, limitations, implications):

    • Spawn argument-builder for theoretical framing guidance
    • The agent can draft scope conditions using the theory section's literatures
  2. Evidence issues (claims need more support, alternative explanations):

    • Qualitative projects: Spawn interview-analyst Phase 2-3 to find additional evidence, then qual-findings-writer Phase 3
    • Quantitative projects: Spawn r-analyst or stata-analyst Phase 3-4 for additional analysis, then quant-findings-writer Phase 5
    • Mixed methods projects: May need to return to either or both strands, then mixed-methods-findings-writer Phase 5

Never fabricate scope conditions, limitations, or statistical results. Route to the appropriate skill.

What You Need

  1. The manuscript (complete draft or relevant sections)
  2. The feedback (any format: bulleted, prose, structured)
  3. Supporting materials (if available):
    • Original research question and argument
    • Data/analysis files
    • Prior versions (for tracking changes)

File Management

This skill uses git to track progress across phases. Before modifying any output file at a new phase:

  1. Stage and commit current state: git add [files] && git commit -m "revision-coordinator: Phase N complete"
  2. Then proceed with modifications.

Do NOT create version-suffixed copies (e.g., -v2, -final, -working). The git history serves as the version trail.

Core Principles

  1. Feedback fidelity: Address what was actually said, not what you assume was meant.

  2. Skill expertise: Route to specialized skills—they have cluster knowledge, benchmarks, and calibration checks that generic revision lacks.

  3. Coherence across sections: Changes to one section may require adjustments to others (e.g., intro changes may break conclusion callbacks).

  4. Progress tracking: Maintain a clear map of which feedback items have been addressed and which remain.

  5. Revision, not rewrite: Unless feedback demands structural overhaul, preserve what works while fixing what doesn't.

Workflow Phases

Phase 0: Intake & Feedback Mapping

Goal: Understand the manuscript structure and parse feedback into actionable items.

Process:

  • Read the full manuscript (or available sections)
  • Read the feedback carefully
  • Parse feedback into discrete items (one issue per item)
  • Categorize each item by type:
    • Structural: Architecture, organization, missing sections
    • Substantive: Argument, evidence, interpretation
    • Methodological: Methods justification, credibility, transparency
    • Stylistic: Word count, repetition, clarity
    • Coherence: Cross-section alignment, promise-delivery match
  • Map each item to the section it addresses
  • Identify which skills are relevant for each section
  • Create the Revision Task List

Output: revision-map.md with parsed feedback and skill assignments.

Pause: User confirms feedback parsing and skill routing.


Phase 1: Diagnostic Assessment

Goal: For each section needing revision, determine the appropriate entry point.

Process:

  • For each section mapped to a specialized skill:
    • Identify the relevant cluster/pathway (using skill's Phase 0 logic)
    • Assess current draft against cluster benchmarks
    • Determine issue severity (minor calibration vs. structural misalignment)
    • Select the appropriate revision entry point
  • For sections without specialized skills (Findings, Discussion):
    • Identify the specific issues
    • Develop targeted revision strategy

Output: Updated revision-map.md with diagnostic findings and entry points.

Pause: User confirms diagnostic assessment and revision strategy.


Phase 1.5 (Optional): Pre-Empirical Peer Review

Goal: Validate theoretical framing before diving deep into empirical analysis.

When to Use:

  • Before committing to a major empirical deep dive
  • When theoretical framing is new or unconventional
  • When you want to stress-test the theory section before it shapes findings analysis
  • When substantial theory revisions have been made and you want external validation

Process: Invoke the peer-reviewer skill to simulate peer review of your theoretical framework:

  1. Scope the review: Focus on Introduction + Theory sections (not full manuscript)
  2. Configure reviewers: Request 2-3 reviewer personas relevant to your theoretical traditions
  3. Target questions:
    • Is the gap clearly articulated?
    • Is the theoretical contribution distinct from prior work?
    • Are the right literatures engaged?
    • Does the framing set up the empirical work appropriately?

Invocation:

Task: Pre-Empirical Peer Review
subagent_type: general-purpose
model: opus
prompt: |
  Load the peer-reviewer skill (read ../peer-reviewer/SKILL.md).

  SCOPE: Theory validation review (Introduction + Theory only)

  MANUSCRIPT SECTIONS:
  [paste Introduction and Theory sections]

  FOCUS QUESTIONS:
  1. Is the theoretical gap clearly defined?
  2. Does the proposed contribution address the gap?
  3. Are the engaged literatures appropriate and sufficient?
  4. Does this theoretical framing set up empirical analysis effectively?

  Run Phases 1-4 (stop before revision support). Request 2-3 reviewer personas
  grounded in the theoretical traditions engaged in this work.

Output: Reviewer feedback on theoretical framing + synthesis of major concerns.

Decision Point: Based on peer-reviewer feedback:

  • If theoretical framing is solid → proceed to Phase 2 (Skill Dispatch)
  • If major concerns identified → route back to Phase 1 diagnostics with new issues
  • If minor calibration needed → incorporate feedback into Phase 2 routing

Pause: User reviews simulated peer feedback and decides whether to proceed or revise theory first.


Phase 2: Skill Dispatch via Task Agents

Goal: Spawn Task agents to revise each section. You coordinate; agents do the work.

MANDATORY: All revision happens through Task agents.

You do NOT revise text yourself. For each section:

  1. Prepare the dispatch context (see Dispatch Templates below)
  2. Spawn a Task agent with the appropriate skill
  3. Wait for agent completion
  4. Update revision-map.md with results
  5. Proceed to next section

Dispatch Protocol for Each Section:

When spawning a Task agent, provide:

  1. Skill path: Which skill.md to load
  2. Phase specification: Which phase of that skill to run
  3. The existing section text (what needs revision)
  4. The specific feedback items (what needs to change)
  5. The identified cluster/pathway (from diagnostic)
  6. The contextual sections (intro-bookends needs Theory+Findings; argument-builder needs RQ+argument)
  7. Output location: Where to save the revised section

Tracking: Mark each feedback item as:

  • [ ] Pending
  • [~] In progress (agent spawned)
  • [x] Addressed (agent completed)
  • [!] Needs user decision

Output: Agent-revised sections + updated tracking in revision-map.md.

Pause after each major section: User reviews agent's revisions before proceeding.


Phase 3: Integration Review

Goal: Ensure revisions are coherent across the manuscript.

Cross-Section Checks:

  • Intro → Findings/Discussion: Do intro promises match what's delivered?
  • Theory → Findings: Do theoretical concepts appear in findings analysis?
  • Methods → Findings: Do methods support the claims made?
  • Intro → Conclusion: Are there callbacks? Does the conclusion answer the intro's question?
  • Terminology: Is key language consistent throughout?
  • Citation: Are sources cited consistently?

Coherence Repairs:

  • If intro promises changed, may need to adjust conclusion
  • If theory framing changed, may need to revise findings language
  • Use article-bookends Phase 3 for intro/conclusion coherence specifically

Output: Coherence assessment + any final adjustments.

Pause: User confirms cross-section coherence.


Phase 4: Verification & Response

Goal: Confirm all feedback addressed and prepare revision summary.

Process:

  • Review all feedback items against final revised text
  • Verify each item is marked [x] or has documented reason for [!]
  • Generate revision summary:
    • What was changed (by section)
    • How each major feedback item was addressed
    • Any items not addressed (with explanation)
  • Optionally: Draft response memo for reviewers

Output: Updated revision-map.md with complete feedback accounting; response-memo.md if R&R.

Recommended: Before final submission, run /writing-editor on the complete manuscript for prose polish—fixes passive voice, abstract nouns, throat-clearing, and academic bad habits using a top-down (Document → Paragraph → Sentence → Word) workflow with human checkpoints.


Phase 5 (Optional): Final Peer Review Simulation

Goal: Simulate peer review of the complete manuscript before submission.

When to Use:

  • After all revisions are complete and integrated
  • Before submitting to a journal
  • When you want to anticipate likely reviewer concerns
  • When targeting a specific journal and want discipline-appropriate feedback

Process: Invoke the peer-reviewer skill for full manuscript review:

  1. Full manuscript scope: Provide the complete revised manuscript
  2. Configure reviewers: Request 3-4 reviewer personas representing:
    • Your primary theoretical tradition
    • Your methodological approach
    • A skeptical "generalist" perspective
    • (Optional) A specific journal's likely reviewer pool
  3. Standard review criteria:
    • Contribution clarity and significance
    • Theoretical engagement and positioning
    • Methodological rigor and transparency
    • Evidence quality and interpretation
    • Writing quality and organization

Invocation:

Task: Final Peer Review Simulation
subagent_type: general-purpose
model: opus
prompt: |
  Load the peer-reviewer skill (read ../peer-reviewer/SKILL.md).

  SCOPE: Full manuscript peer review simulation

  MANUSCRIPT:
  [paste complete manuscript OR provide file path]

  TARGET JOURNAL (if applicable): [journal name]

  REVIEWER CONFIGURATION:
  - Primary theoretical tradition: [e.g., social movement studies]
  - Methodological approach: [e.g., qualitative/interview-based]
  - Include one generalist/skeptical reviewer

  Run all phases (1-6). Generate substantive reviews that identify:
  1. Major concerns (potential reject/revise-and-resubmit issues)
  2. Minor concerns (addressable in revision)
  3. Strengths to preserve

  Conclude with synthesis and prioritized revision recommendations.

Output:

  • 3-4 simulated reviewer reports
  • Synthesis of overlapping concerns
  • Prioritized list of pre-submission revisions
  • (Optional) Draft response strategy for anticipated concerns

Decision Point: Based on simulated reviews:

  • If reviews suggest "accept with minor revisions" → manuscript ready for submission
  • If reviews suggest "revise and resubmit" → return to Phase 0 with new feedback
  • If reviews identify fundamental issues → major revision cycle needed

Pause: User reviews simulated peer feedback and decides on next steps.


Folder Structure for Revision

project/
├── manuscript/
│   ├── first-draft.md           # Original manuscript (edited in place by dispatched agents)
│   └── feedback.md              # Reviewer/editor feedback
├── revision/
│   ├── revision-map.md          # Central tracking doc (grows across phases; git tracks)
│   └── response-memo.md         # R&R deliverable (Phase 4, if applicable)
├── peer-review/                 # Peer-reviewer skill outputs
│   ├── pre-empirical-review.md  # Phase 1.5 output (if used)
│   ├── reviewer-personas/       # Generated reviewer profiles
│   ├── reviews/                 # Individual simulated reviews
│   └── final-review-synthesis.md # Phase 5 output (if used)

Diagnostics are presented in conversation. Section revisions are made in place in the manuscript; git tracks each phase's state. No diagnostics/, section-revisions/, or revised-draft.md files are created.

Feedback Parsing Guidelines

Parse Into Discrete Items

Transform this:

"The intro is too long and repetitive—you have two intros. Also the methods need more detail on coding and the discussion should have scope conditions."

Into:

1. [Intro] Length: Intro too long
2. [Intro] Structure: Two intros detected (repetition)
3. [Methods] Credibility: More detail on coding needed
4. [Discussion] Scope: Add scope conditions

Categorize by Type

Type Examples Typical Skill Response
Structural "Reorganize the theory section" Skill Phase 1 (Architecture)
Substantive "Strengthen the argument for X" Skill Phase 3-4 (Drafting/Turn)
Methodological "Explain intercoder reliability" methods-writer Phase 2
Stylistic "Cut 500 words from intro" Skill calibration checks
Coherence "Intro promises don't match findings" article-bookends Phase 3
Statistical "Re-run models with controls for X" r-analyst/stata-analyst Phase 3-4
Integration "Better connect qual and quant" mixed-methods-findings-writer Phase 2-5

Identify Dependencies

Some feedback items depend on others:

  • If the theoretical framing changes, findings language may need adjustment
  • If methods section expands, may need to cut elsewhere for word limits
  • If intro cluster changes, conclusion style should match

Note dependencies in the revision map so sequencing is correct.

Invoking Peer-Reviewer Skill

The peer-reviewer skill simulates peer review using local library materials to construct informed reviewer personas. Unlike other sub-skills that revise specific sections, peer-reviewer provides external validation at strategic checkpoints.

Key Features:

  • Builds reviewer personas from your local library (named for theoretical perspectives, not individual scholars)
  • Uses full-text retrieval to ground feedback in actual disciplinary knowledge
  • Returns structured reviews with major concerns, minor concerns, and strengths
  • Provides synthesis and prioritized revision recommendations

When to Invoke:

Checkpoint Phase Purpose
Pre-empirical Phase 1.5 Validate theoretical framing before data analysis
Pre-submission Phase 5 Anticipate reviewer concerns before journal submission

Prerequisites:

  • references.bib and library/ folder accessible
  • Relevant literature in your local library (peer-reviewer will search for works related to your theoretical traditions)
  • Clear articulation of your theoretical positioning

Dispatch Templates (MANDATORY)

You MUST use Task agents for all revision work. Copy and customize these templates.


Theory Section Dispatch

Task: Revise Theory Section
subagent_type: general-purpose
model: opus
prompt: |
  Load the argument-builder skill:
  - Read: [path]/argument-builder/SKILL.md
  - Read: [path]/argument-builder/phases/phase5-revision.md
  - Read: [path]/argument-builder/clusters/[cluster-name].md

  TASK: Revise an existing Theory section (not drafting fresh).

  EXISTING SECTION:
  [paste current theory section - FULL TEXT]

  CLUSTER IDENTIFIED: [Gap-Filler / Theory-Extender / Concept-Builder / Synthesis / Problem-Driven]

  FEEDBACK TO ADDRESS:
  1. [ID: T-1] [specific feedback item]
  2. [ID: T-2] [specific feedback item]

  CONTEXT:
  - Research question: [RQ]
  - Main argument: [argument]
  - Key literatures: [list]

  CALIBRATION TARGETS:
  - Word count: 1145-1744 (median 1525)
  - Citation density: ~24 per 1000 words

  Run Phase 5 (Revision) calibration checks. Revise the section to address
  the feedback while maintaining [Cluster] characteristics.

  OUTPUT:
  - Edit the theory section in place in the manuscript file. Git commit after.
  - Return: summary of changes, items addressed, remaining issues

Findings Section Dispatch

Task: Revise Findings Section
subagent_type: general-purpose
model: opus
prompt: |
  Load the qual-findings-writer skill:
  - Read: [path]/qual-findings-writer/SKILL.md
  - Read: [path]/qual-findings-writer/phases/phase3-revision.md
  - Read: [path]/qual-findings-writer/techniques/rubric.md

  TASK: Revise an existing Findings section using Phase 3 revision protocol.

  EXISTING SECTION:
  [paste current findings section - FULL TEXT]

  FEEDBACK TO ADDRESS:
  1. [ID: F-1] [specific feedback item]
  2. [ID: F-2] [specific feedback item]

  CONTEXT:
  - Research question: [RQ]
  - Main argument: [argument]
  - Theoretical concepts (from theory section): [list key concepts that should appear]

  DATA ACCESS:
  - Quote database: [path to analysis/quote-database.md]
  - Participant profiles: [path to analysis/participant-profiles/]
  - Interview transcripts: [path to interviews/]

  Run Phase 3 revision checks:
  1. Argument structure (roadmap, claim-first subsections)
  2. Anchor-echo pattern verification
  3. Quote integration quality
  4. Voice and confidence
  5. Variation handling
  6. Scope and prevalence indicators

  If additional evidence is needed that isn't in the quote database,
  note this as "NEEDS DATA RETURN" - do not fabricate quotes.

  OUTPUT:
  - Edit the findings section in place in the manuscript file. Git commit after.
  - Return: summary of changes, items addressed, any NEEDS DATA RETURN flags

Introduction Dispatch

Task: Revise Introduction
subagent_type: general-purpose
model: opus
prompt: |
  Load the article-bookends skill:
  - Read: [path]/article-bookends/SKILL.md
  - Read: [path]/article-bookends/phases/phase1-introduction.md
  - Read: [path]/article-bookends/clusters/[cluster-name].md

  TASK: Revise an existing Introduction.

  EXISTING SECTION:
  [paste current introduction - FULL TEXT]

  CLUSTER IDENTIFIED: [Gap-Filler / Theory-Extension / Concept-Building / Synthesis / Problem-Driven]

  FEEDBACK TO ADDRESS:
  1. [ID: I-1] [specific feedback item]
  2. [ID: I-2] [specific feedback item]

  CONTEXT FOR COHERENCE:
  - Theory section summary: [key argument and contribution]
  - Findings delivered: [what the findings actually show]
  - Current conclusion: [for callback alignment]

  TARGET: ~761 words, 6 paragraphs (median)

  Revise to address feedback while ensuring:
  - Opening move matches [Cluster] type
  - Promises align with what findings actually deliver
  - Gap/puzzle is clearly established
  - Roadmap previews actual structure

  OUTPUT:
  - Edit the introduction in place in the manuscript file. Git commit after.
  - Return: summary of changes, opening move type used

Conclusion Dispatch

Task: Revise Conclusion
subagent_type: general-purpose
model: opus
prompt: |
  Load the article-bookends skill:
  - Read: [path]/article-bookends/SKILL.md
  - Read: [path]/article-bookends/phases/phase2-conclusion.md
  - Read: [path]/article-bookends/techniques/callbacks.md

  TASK: Revise an existing Conclusion.

  EXISTING SECTION:
  [paste current conclusion - FULL TEXT]

  CLUSTER IDENTIFIED: [Gap-Filler / Theory-Extension / Concept-Building / Synthesis / Problem-Driven]

  FEEDBACK TO ADDRESS:
  1. [ID: C-1] [specific feedback item]

  CONTEXT FOR COHERENCE:
  - Introduction (for callbacks): [paste current introduction]
  - Key findings: [list main findings]
  - Contribution claim: [the paper's core contribution]

  Revise to address feedback while ensuring:
  - Callbacks to introduction language are present
  - Contribution is explicit and specific
  - Implications are grounded in findings
  - Matches [Cluster] conclusion style

  OUTPUT:
  - Edit the conclusion in place in the manuscript file. Git commit after.
  - Return: summary of changes, callback phrases identified

Methods Section Dispatch

Task: Revise Methods Section
subagent_type: general-purpose
model: opus
prompt: |
  Load the methods-writer skill:
  - Read: [path]/methods-writer/SKILL.md
  - Read: [path]/methods-writer/phases/phase2-revision.md
  - Read: [path]/methods-writer/pathways/[pathway-name].md

  TASK: Revise an existing Methods section.

  EXISTING SECTION:
  [paste current methods section - FULL TEXT]

  PATHWAY IDENTIFIED: [Efficient / Standard / Detailed]

  FEEDBACK TO ADDRESS:
  1. [ID: M-1] [specific feedback item]
  2. [ID: M-2] [specific feedback item]

  STUDY DETAILS:
  - Sample size: [N]
  - Population: [description]
  - Recruitment: [approach]
  - Interview format: [semi-structured, etc.]
  - Analysis approach: [method]

  PATHWAY TRIGGERS:
  - Vulnerable population? [Yes/No]
  - Novel methods? [Yes/No]
  - Space constraints? [Yes/No]

  If pathway should change based on triggers, redraft accordingly.

  OUTPUT:
  - Edit the methods section in place in the manuscript file. Git commit after.
  - Return: pathway used, word count, components added/removed

Case Justification Dispatch

Task: Revise Case Justification
subagent_type: general-purpose
model: opus
prompt: |
  Load the case-justification skill:
  - Read: [path]/case-justification/SKILL.md
  - Read: [path]/case-justification/phases/phase2-revision.md
  - Read: [path]/case-justification/clusters/[cluster-name].md

  TASK: Revise an existing Case Justification section.

  EXISTING SECTION:
  [paste current case section - FULL TEXT]

  CLUSTER IDENTIFIED: [Minimal / Standard / Deep Historical / Comparative / Policy-Driven]

  FEEDBACK TO ADDRESS:
  1. [ID: CJ-1] [specific feedback item]

  CASE DETAILS:
  - Site(s): [description]
  - Population: [description]
  - Key contextual features: [list]
  - Relationship to theory: [why this case for this argument?]

  POSITION CHECK:
  - Current position: [Before/After theory section]
  - Required position for cluster: [Policy-Driven = before; others = after]

  OUTPUT:
  - Edit the case justification in place in the manuscript file. Git commit after.
  - Return: cluster used, word count, position recommendation

Discussion Dispatch (Scope/Framing Issues)

Task: Revise Discussion - Scope and Framing
subagent_type: general-purpose
model: opus
prompt: |
  Load the argument-builder skill:
  - Read: [path]/argument-builder/SKILL.md
  - Read: [path]/argument-builder/techniques/scope-conditions.md (if exists)

  TASK: Revise Discussion section for scope conditions, limitations, implications.

  EXISTING SECTION:
  [paste current discussion section - FULL TEXT]

  FEEDBACK TO ADDRESS:
  1. [ID: D-1] [specific feedback item about scope/limitations/implications]

  CONTEXT:
  - Theory section literatures: [list engaged literatures]
  - Contribution claim: [the paper's core contribution]
  - Key findings: [list main findings]

  DISCUSSION COMPONENTS TO CHECK:
  - Summary: Brief recap (not repetition)
  - Implications: How findings advance literature
  - Scope conditions: Explicit boundaries on claims
  - Limitations: Honest but not self-undermining
  - Future directions: 2-3 specific suggestions

  Revise to address feedback. Ground scope conditions in the actual
  study design and sample - do not invent limitations.

  OUTPUT:
  - Edit the discussion section in place in the manuscript file. Git commit after.
  - Return: summary of changes, scope conditions added

Return to Data Dispatch (When More Evidence Needed)

Use this when feedback indicates claims need more support or quotes are insufficient.

Task: Return to Data for Additional Evidence
subagent_type: general-purpose
model: opus
prompt: |
  Load the interview-analyst skill:
  - Read: [path]/interview-analyst/SKILL.md
  - Read: [path]/interview-analyst/phases/phase2-coding.md
  - Read: [path]/interview-analyst/phases/phase3-interpretation.md

  TASK: Find additional evidence for a specific claim or theme.

  EVIDENCE NEEDED FOR:
  [describe what claim or section needs more support]

  CURRENT QUOTE DATABASE:
  [path to analysis/quote-database.md]

  INTERVIEW TRANSCRIPTS:
  [path to interviews/]

  EXISTING CODES RELEVANT TO THIS CLAIM:
  [list relevant codes from codebook if available]

  PROCESS:
  1. Search transcripts for quotes supporting this claim
  2. Check if existing codes capture this or if new coding needed
  3. Add new quotes to quote database with proper tagging
  4. Note prevalence (how many participants?)

  OUTPUT:
  - Update: analysis/quote-database.md with new quotes
  - Return: quotes found, prevalence count, any new codes created
  - Flag if insufficient evidence exists in the data

Quantitative Results Section Dispatch

Task: Revise Quantitative Results Section
subagent_type: general-purpose
model: opus
prompt: |
  Load the quant-findings-writer skill:
  - Read: [path]/quant-findings-writer/SKILL.md
  - Read: [path]/quant-findings-writer/clusters/[cluster-name].md
  - Read: [path]/quant-findings-writer/techniques/techniques.md

  TASK: Revise an existing Results section.

  EXISTING SECTION:
  [paste current results section - FULL TEXT]

  CLUSTER IDENTIFIED: [Progressive Model Builder / Hypothesis Tester / Decomposition Analyst / etc.]

  FEEDBACK TO ADDRESS:
  1. [ID: R-1] [specific feedback item]
  2. [ID: R-2] [specific feedback item]

  CONTEXT:
  - Research question: [RQ]
  - Main argument: [contribution]
  - Statistical output location: [path to tables/figures]

  STATISTICAL OUTPUT:
  - Tables: [list key tables]
  - Figures: [list key figures]
  - Model documentation: [path to analysis scripts/output]

  Run Phase 5 (Revision) calibration checks. Revise to address feedback
  while maintaining [Cluster] characteristics.

  If reanalysis is needed (new models, different specifications),
  note this as "NEEDS REANALYSIS" - do not fabricate statistical results.

  OUTPUT:
  - Edit the results section in place in the manuscript file. Git commit after.
  - Return: summary of changes, items addressed, any NEEDS REANALYSIS flags

Mixed Methods Findings Section Dispatch

Task: Revise Mixed Methods Findings Section
subagent_type: general-purpose
model: opus
prompt: |
  Load the mixed-methods-findings-writer skill:
  - Read: [path]/mixed-methods-findings-writer/SKILL.md
  - Read: [path]/mixed-methods-findings-writer/clusters/[cluster-name].md
  - Read: [path]/mixed-methods-findings-writer/techniques/techniques.md

  TASK: Revise an existing integrated Findings section.

  EXISTING SECTION:
  [paste current findings section - FULL TEXT]

  CLUSTER IDENTIFIED: [Thematic Integrator / Quant-Anchored Elaborator / etc.]

  FEEDBACK TO ADDRESS:
  1. [ID: MF-1] [specific feedback item]
  2. [ID: MF-2] [specific feedback item]

  CONTEXT:
  - Research question: [RQ]
  - Main argument: [contribution]
  - Integration design: [how qual and quant connect]

  QUALITATIVE DATA ACCESS:
  - Quote database: [path]
  - Interview transcripts: [path]

  QUANTITATIVE DATA ACCESS:
  - Statistical output: [path to tables/figures]
  - Model documentation: [path]

  Run Phase 5 (Revision). Revise to address feedback while maintaining
  integration coherence between qualitative and quantitative strands.

  If additional qualitative evidence is needed, note as "NEEDS QUAL DATA RETURN".
  If reanalysis is needed, note as "NEEDS REANALYSIS".
  Do not fabricate quotes or statistical results.

  OUTPUT:
  - Edit the mixed methods findings section in place in the manuscript file. Git commit after.
  - Return: summary of changes, integration assessment, any data return flags

Return to Statistical Analysis Dispatch (When Reanalysis Needed)

Use this when feedback indicates statistical models need revision or additional analysis.

Task: Return to Statistical Analysis for Reanalysis
subagent_type: general-purpose
model: opus
prompt: |
  Load the r-analyst skill (or stata-analyst as appropriate):
  - Read: [path]/r-analyst/SKILL.md
  - Read: [path]/r-analyst/phases/phase3-estimation.md
  - Read: [path]/r-analyst/phases/phase4-robustness.md

  TASK: Revise statistical analysis based on reviewer feedback.

  REANALYSIS NEEDED FOR:
  [describe what models or analyses need revision]

  CURRENT ANALYSIS:
  - Scripts: [path to analysis scripts]
  - Output: [path to tables/figures]
  - Data: [path to processed data]

  FEEDBACK DRIVING REANALYSIS:
  1. [specific feedback about statistical approach]

  PROCESS:
  1. Review current model specification
  2. Implement requested changes (alternative estimator, additional controls, subgroup analysis, etc.)
  3. Re-run robustness checks (Phase 4)
  4. Update output documentation (Phase 5)
  5. Regenerate tables and figures

  OUTPUT:
  - Updated analysis scripts
  - Updated tables/figures
  - Return: summary of changes, new results, any concerns about revised specification

Coherence Check Dispatch

Task: Coherence Check - Intro/Conclusion Alignment
subagent_type: general-purpose
model: opus
prompt: |
  Load the article-bookends skill:
  - Read: [path]/article-bookends/SKILL.md
  - Read: [path]/article-bookends/phases/phase3-coherence.md
  - Read: [path]/article-bookends/techniques/callbacks.md

  TASK: Verify and repair coherence between Introduction and Conclusion.

  INTRODUCTION:
  [paste current introduction - FULL TEXT]

  CONCLUSION:
  [paste current conclusion - FULL TEXT]

  THEORY SECTION (for reference):
  [paste or summarize theory section]

  FINDINGS SECTION (for reference):
  [paste or summarize key findings]

  COHERENCE CHECKS:
  1. Vocabulary echoes (key terms in both)
  2. Promise-delivery alignment (intro promises match findings)
  3. Callback presence (conclusion references intro language)
  4. Ambition calibration (claims consistent across sections)

  Classify coherence type:
  - Parallel (66%): Promises match delivery ✓
  - Escalators (20%): Delivery exceeds promises (acceptable)
  - Deflators (6%): Delivery falls short (problematic - fix!)

  OUTPUT:
  - Present coherence assessment in conversation; add findings to `revision/revision-map.md`.
  - If fixes needed, edit sections in place in the manuscript file. Git commit after.
  - Return: coherence type, issues found, fixes made

Abstract Revision Dispatch

Task: Revise Abstract
subagent_type: general-purpose
model: opus
prompt: |
  Load the abstract-builder skill:
  - Read: [path]/abstract-builder/SKILL.md
  - Read: [path]/abstract-builder/phases/phase3-revision.md
  - Read: [path]/abstract-builder/clusters/[archetype-name].md

  TASK: Revise an existing abstract.

  EXISTING ABSTRACT:
  [paste current abstract - FULL TEXT]

  ARCHETYPE IDENTIFIED: [Empirical-Showcase / Research-Report / Stakes-Driven / Puzzle-Solver]

  FEEDBACK TO ADDRESS:
  1. [ID: A-1] [specific feedback item]

  CONTEXT:
  - Research question: [RQ]
  - Main argument: [contribution]
  - Key findings: [2-3 findings]
  - Data: [sample size, population, location]

  CALIBRATION TARGETS:
  - Word count: 165-210 (median 189)
  - Sentence count: 5-7 (median 6)
  - Closing: Should end with contribution claim

  Generate 2-3 variants using different archetypes for comparison.

  OUTPUT:
  - Present 2-3 variants in conversation for user to select one, then edit the abstract in place in the manuscript file. Git commit after.
  - Return: recommended variant, word counts, archetype comparison

Additional Literature Search Dispatch

Use when reviewer asks for engagement with literatures not covered.

Task: Expand Literature Coverage
subagent_type: general-purpose
model: sonnet
prompt: |
  Load the lit-search skill:
  - Read: [path]/lit-search/SKILL.md
  - Read: [path]/lit-search/api/openalex-reference.md

  TASK: Find additional literature on a specific topic/tradition.

  LITERATURES TO SEARCH:
  [describe what literatures reviewer asked for - be specific]

  EXISTING COVERAGE:
  [list literatures already in the theory section]

  SEARCH CONSTRAINTS:
  - Date range: [if specified]
  - Key authors to include: [if known]
  - Disciplinary focus: [sociology, etc.]

  PROCESS:
  1. Develop search terms for the missing literature
  2. Run OpenAlex queries
  3. Screen for relevance (focus on seminal works + recent advances)
  4. Return 5-10 key papers that should be engaged

  OUTPUT:
  - Save to: literature/additional-search-[topic].json
  - Return: list of papers with brief relevance notes, suggested citation pattern

Literature Synthesis Dispatch

Use when reviewer asks for deeper engagement with existing literature or theoretical positioning.

Task: Deepen Literature Engagement
subagent_type: general-purpose
model: opus
prompt: |
  Load the lit-synthesis skill:
  - Read: [path]/lit-synthesis/SKILL.md
  - Read: [path]/lit-synthesis/phases/phase4-debate-mapping.md

  TASK: Develop deeper engagement with specific theoretical traditions.

  FEEDBACK DRIVING THIS:
  [describe what reviewer asked for - more engagement, better positioning, etc.]

  LITERATURES TO ENGAGE:
  [list specific literatures or debates]

  EXISTING LITERATURE DATABASE:
  - references.bib: [path]
  - Library folder: [path to library/ folder]

  CURRENT THEORY SECTION (for context):
  [paste or summarize current theoretical framing]

  PROCESS:
  1. Identify key debates in the specified literatures
  2. Map positions and tensions
  3. Locate where this paper's contribution fits
  4. Suggest specific citations and framing moves

  OUTPUT:
  - Save to: literature/synthesis-[topic].md
  - Return: debate map, positioning recommendations, suggested citations

Writing Editor Dispatch (Prose Polish)

Use at the END of revision, after all substantive changes are complete.

Task: Polish Prose - Full Manuscript
subagent_type: general-purpose
model: opus
prompt: |
  Load the writing-editor skill:
  - Read: [path]/writing-editor/SKILL.md
  - Read: [path]/writing-editor/references/merged-guidelines.md

  TASK: Edit the complete manuscript for prose quality.

  MANUSCRIPT:
  [paste complete manuscript OR provide file path]

  PROTECTION MODE: [copy / git]

  RUN THE FOUR-LEVEL WORKFLOW:
  1. Document level: hooks, titles, structure, citations
  2. Paragraph level: endings, symmetry, triplets, contrast
  3. Sentence level: passive voice, abstract nouns, agents, dashes
  4. Word level: throat-clearing, signposts, adverbs, fancy-talk

  CONSTRAINTS:
  - Preserve all citations exactly
  - Maintain academic register
  - Do not change substantive claims

  OUTPUT:
  - Save edited version to: [manuscript-edited.md or via git commit]
  - Return: summary of changes at each level
  - Pause for user approval between levels

Writing Editor Dispatch (Single Section)

Use for prose polish on a specific section.

Task: Polish Prose - [Section Name]
subagent_type: general-purpose
model: opus
prompt: |
  Load the writing-editor skill:
  - Read: [path]/writing-editor/SKILL.md
  - Read: [path]/writing-editor/references/merged-guidelines.md

  TASK: Edit a single section for prose quality.

  SECTION:
  [paste section text - FULL TEXT]

  SECTION TYPE: [Introduction / Theory / Methods / Findings / Discussion / Conclusion]

  RUN ALL FOUR LEVELS on this section:
  1. Document level (for this section): structure, flow
  2. Paragraph level: endings, symmetry
  3. Sentence level: passive voice, abstract nouns
  4. Word level: throat-clearing, signposts

  CONSTRAINTS:
  - Preserve all citations exactly
  - Maintain academic register
  - Keep substantive claims unchanged

  OUTPUT:
  - Edit the section in place in the manuscript file. Git commit after.
  - Return: summary of changes, word count before/after

Common Revision Scenarios

Scenario: "Two Intros" Problem

Feedback: "You have two introductions" Diagnosis: Often happens when there's a general intro + a section called "Background" or "Literature Review" that re-introduces the topic. Resolution:

  1. Keep ONE intro (usually the first)
  2. Convert the second into a proper Theory section (use argument-builder cluster guidance)
  3. Run article-bookends Phase 3 for coherence check

Scenario: Methods Credibility Gap

Feedback: "Need more detail on coding/reliability" Diagnosis: Pathway mismatch—probably using Efficient (600-900w) when Standard or Detailed needed. Resolution:

  1. Re-run methods-writer Phase 0 to confirm pathway
  2. If pathway should change, redraft with new word allocation
  3. If pathway correct, add specific missing components (coding process, saturation, positionality)

Scenario: Weak Turn in Theory

Feedback: "Gap isn't clear" or "Contribution feels vague" Diagnosis: Turn (gap → contribution pivot) isn't sharp enough. Resolution:

  1. Use argument-builder Phase 4 (Turn) specifically
  2. Ensure turn appears at subsection transition, not buried
  3. Check that "what we don't know" is specific, not generic

Scenario: Promise-Delivery Mismatch

Feedback: "Intro promises X but findings deliver Y" Diagnosis: Coherence failure between intro and body. Resolution:

  1. Decide which is right: the promise or the delivery
  2. If delivery is right, revise intro to match (article-bookends Phase 1)
  3. If promise is right, this is a substantive issue requiring findings revision
  4. Run article-bookends Phase 3 for coherence verification

Key Reminders

  • Don't over-revise: Fix what feedback identifies; preserve what works.
  • Track everything: The revision map is your accountability document.
  • Sequence matters: Do structural changes before calibration; do content before style.
  • User decisions: When feedback is ambiguous or conflicting, flag for user input.
  • Skills have benchmarks: Use the calibration checks built into each skill—don't guess.
  • Coherence is a system property: Changes to one section affect others.
  • Peer review strategically: Use Phase 1.5 (pre-empirical) when theory is novel or risky; use Phase 5 (final) when manuscript is complete and ready for submission feedback. The peer-reviewer skill uses local library materials to build informed reviewer personas—it's most valuable when you want discipline-specific critique.

Starting the Process

When the user is ready to begin:

  1. Ask for the manuscript:

    "Please share your manuscript (or the sections you want revised). I need to see the current state."

  2. Ask for the feedback:

    "Please share the feedback you've received. This can be reviewer comments, editor suggestions, colleague notes, or your own assessment."

  3. Ask about priorities:

    "Is there anything you specifically agree or disagree with in the feedback? Any constraints (word limits, sections that can't change, etc.)?"

  4. Proceed with Phase 0 to parse and map the feedback.

Weekly Installs
9
GitHub Stars
3
First Seen
14 days ago
Installed on
opencode9
gemini-cli9
claude-code9
github-copilot9
codex9
amp9