qa-management
Q&A Management
As transactions enter confirmatory diligence, the Q&A process can devolve into a logistical nightmare. Multiple bidders submit hundreds of granular questions about tax liabilities, employee benefits, and IT infrastructure. Junior deal team members become administrative traffic cops, routing questions, chasing answers, and formatting responses.
When to Use
- Responding to buyer diligence question trackers
- Managing multi-bidder Q&A in a sell-side auction
- Ensuring consistency in seller representations across bidders
- Drafting responses using approved VDR documents and disclosure schedules
The /draft-qna Protocol
How It Works
- Connect to approved VDR documents, existing disclosure schedules, and prior Q&A
- Analyze each buyer question to identify the relevant data source
- Draft a preliminary response synthesizing approved materials
- Cross-reference against previously answered questions for consistency
- Flag for expert approval with routing recommendations
- Track response status across all bidders
Output Format
## Q&A Response Draft
### Question [#]
**From**: [Bidder name]
**Category**: [Financial / Legal / HR / Operations / IT / Tax]
**Question**: [Full question text]
**Draft Response**:
[Response synthesized from approved materials]
**Sources Used**:
- [VDR document reference and folder path]
- [Prior Q&A reference if applicable]
**Routing**: → [CFO / General Counsel / CTO] for approval
**Sensitivity**: [Standard / Restricted / Highly Confidential]
**Consistency Check**: [Aligned with prior responses / POTENTIAL CONFLICT with Q#X]
Multi-Bidder Consistency Tracking
## Q&A Consistency Matrix
| Topic | Bidder A Response | Bidder B Response | Consistent? |
|-------|------------------|------------------|-------------|
| EBITDA adjustments | Referenced QoE §3.2 | Referenced QoE §3.2 | ✅ |
| Customer concentration | Disclosed top 10 = 45% | Disclosed top 5 = 38% | ⚠️ REVIEW |
| Pending litigation | Full disclosure | Full disclosure | ✅ |
Operating Guidelines
- Never disclose information beyond what's been approved for the current diligence phase
- Maintain strict information barriers between bidders in a competitive auction
- Flag any question that could require disclosure of material non-public information
- Cross-reference every response against prior answers to prevent contradictions
- Route tax questions to tax counsel, employment questions to labor counsel
- Track response turnaround time — delays signal deal fatigue to bidders
- Preserve the Q&A log as it becomes part of the deal's legal record
Time saved: 30 minutes per query, preventing coordination bottlenecks and ensuring consistency.
Examples
Input: "Draft a response to buyer question: 'Explain the $3.2M revenue reclassification in Q3 2023.'"
Auto-drafted response:
The Q3 2023 reclassification reflects adoption of ASC 606 effective July 1, 2023. Previously, installation revenue was recognized at contract signing; under the new standard, it is deferred and recognized over the implementation period (typically 60–90 days). The change had no cash impact. Supporting documentation is available in the VDR at [4.3.2 — Revenue Recognition Policy Memo].
Routed to: CFO for approval. SLA: 24 hours.
Troubleshooting
| Problem | Cause | Fix |
|---|---|---|
| Q&A responses are inconsistent across bidders | No central response register | Maintain a master Q&A log; all answers distributed to all bidders |
| Expert approval is a bottleneck | Routing to wrong reviewer | Build a topic taxonomy; auto-route by category (legal → counsel, financial → CFO) |
| Sensitive information leaked in response | Draft not reviewed | Never send auto-draft directly; stage → review → approve → distribute |
| Response tracker gets out of sync | Manual updates | Use a shared spreadsheet with auto-timestamp on status change |
More from lauraflorentin/skills-marketplace
multi-agent-collaboration
A structural pattern where multiple specialized agents communicate and coordinate to solve a problem that is too complex for a single agent. Use when user asks to "build a multi-agent system", "agents working together", "agent collaboration", or mentions team of agents, distributed agents, or swarm.
21reflection
A recursive pattern where an agent evaluates and critiques its own output to iteratively improve quality and catch errors. Use when user asks to "add self-reflection", "agent introspection", "self-critique", or mentions self-evaluation, meta-cognition, or quality self-assessment.
18human-in-the-loop
A hybrid pattern where the system pauses execution to request human approval, input, or disambiguation before proceeding with critical actions. Use when user asks to "add human approval", "require human review", "human-in-the-loop", or mentions approval workflows, human oversight, or escalation.
16planning
A high-level cognitive pattern where an agent formulates a structured sequence of actions (a plan) before executing any of them, ensuring goal-directed behavior. Use when user asks to "add planning to my agent", "task planning", "agent planning", or mentions plan generation, plan execution, or step-by-step planning.
14parallelization
A concurrency pattern where multiple agent tasks are executed at the same time to speed up processing or gather diverse perspectives. Use when user asks to "run agents in parallel", "parallelize tasks", "concurrent execution", or mentions parallel processing, fan-out, or batch execution.
13routing
A control flow pattern where a central component classifies an input request and directs it to the most appropriate specialized agent or tool. Use when user asks to "route between agents", "agent routing", "task dispatch", or mentions classifier routing, intent detection, or agent selection.
12