deep-research-pro
Deep Research Pro
A professional-grade deep research methodology that coordinates multi-source information gathering, cross-reference verification, and structured synthesis. Designed for research tasks that require high confidence in factual accuracy, comprehensive coverage, and traceable evidence chains.
Overview
Deep Research Pro implements an agent-based methodology where the research process is decomposed into specialized phases: query decomposition, parallel source gathering, cross-reference verification, contradiction resolution, and structured synthesis. Unlike simple search-and-summarize approaches, this skill emphasizes source triangulation, evidence grading, and explicit uncertainty marking.
The methodology is particularly valuable for literature reviews, technology assessments, policy analyses, and any research task where decision-makers need to trust the completeness and accuracy of the findings. Every claim in the final output is linked to at least one verified source, with confidence levels assigned based on the quality and agreement of the evidence.
Research Agent Architecture
Phase 1: Query Decomposition
def decompose_research_query(query: str) -> dict:
"""
Break a complex research question into atomic sub-questions
that can be independently investigated.
Example:
Input: "What is the current state of quantum computing for
drug discovery, and which companies are leading?"
Output: {
'sub_questions': [
'What quantum computing methods are used in drug discovery?',
'What are the computational advantages over classical methods?',
'Which companies are actively working on quantum drug discovery?',
'What drugs or molecules have been studied using quantum methods?',
'What are the current limitations and timeline to practical use?'
],
'cross_cutting_themes': [
'Cost-benefit vs classical computing',
'Regulatory considerations',
'Academic vs industry progress'
],
'source_strategy': {
'academic': ['PubMed', 'arXiv', 'Google Scholar'],
'industry': ['company websites', 'press releases', 'patent databases'],
'grey': ['government reports', 'consulting firm analyses']
}
}
"""
pass # Implemented by the agent's reasoning
Phase 2: Parallel Source Gathering
For each sub-question, gather evidence from multiple independent source types:
| Source Type | Examples | Strengths | Limitations |
|---|---|---|---|
| Peer-reviewed | Journal articles, conference papers | Rigorous review | Publication lag |
| Preprints | arXiv, bioRxiv, SSRN | Current | Not peer-reviewed |
| Institutional | WHO, NIH, government agencies | Authoritative | May be conservative |
| Industry | Company blogs, press releases | Current, practical | Biased toward own products |
| News/Media | Science journalism, trade publications | Accessible | May oversimplify |
| Expert opinion | Interviews, blog posts, talks | Nuanced | Subjective |
Phase 3: Source Verification
def verify_source(source: dict) -> dict:
"""
Assess the credibility and reliability of a single source.
"""
verification = {
'source_id': source['id'],
'url': source['url'],
'checks': {}
}
# Check 1: Author credibility
verification['checks']['author'] = {
'identifiable': bool(source.get('author')),
'affiliated': bool(source.get('institution')),
'h_index_available': bool(source.get('h_index')),
'domain_expert': source.get('expertise_match', False)
}
# Check 2: Publication venue
verification['checks']['venue'] = {
'peer_reviewed': source.get('peer_reviewed', False),
'impact_factor': source.get('impact_factor'),
'known_predatory': source.get('predatory_flag', False)
}
# Check 3: Currency
verification['checks']['currency'] = {
'publication_year': source.get('year'),
'is_recent': source.get('year', 0) >= 2023,
'superseded': source.get('retracted', False)
}
# Check 4: Corroboration
verification['checks']['corroboration'] = {
'cited_by_count': source.get('citation_count', 0),
'independent_confirmation': source.get('replicated', False),
'consistent_with_consensus': source.get('consensus_aligned', None)
}
# Overall confidence score
score = sum([
verification['checks']['author']['identifiable'],
verification['checks']['venue']['peer_reviewed'],
not verification['checks']['venue']['known_predatory'],
verification['checks']['currency']['is_recent'],
verification['checks']['corroboration']['cited_by_count'] > 5
]) / 5.0
verification['confidence'] = round(score, 2)
verification['grade'] = (
'A' if score >= 0.8 else
'B' if score >= 0.6 else
'C' if score >= 0.4 else 'D'
)
return verification
Cross-Reference and Contradiction Resolution
Triangulation Protocol
For each factual claim in the research output:
- Identify the claim: Extract a specific, testable statement.
- Find supporting sources: At least 2 independent sources.
- Check for contradicting sources: Actively search for disconfirming evidence.
- Assess source quality: Weight evidence by source grade (A > B > C > D).
- Assign confidence level: Based on evidence agreement and quality.
Confidence Levels:
HIGH: 3+ Grade A/B sources agree, no contradictions
MEDIUM: 2+ sources agree, minor contradictions or lower-quality sources
LOW: Single source, or significant contradictions among sources
UNCERTAIN: Conflicting evidence of similar quality; state both positions
Contradiction Handling
When sources disagree:
- Check if the disagreement is due to different definitions, timeframes, or contexts.
- Check if one source has been superseded or corrected.
- If genuine disagreement exists, present both positions with their evidence.
- Note the disagreement explicitly: "Source A (Grade B, 2024) reports X, while Source B (Grade A, 2023) reports Y. The discrepancy may be due to [possible explanation]."
Structured Synthesis Output
Report Template
# Research Report: [Topic]
## Executive Summary
[3-5 sentence overview of key findings with confidence levels]
## Methodology
- Sub-questions investigated: [list]
- Sources consulted: [count by type]
- Date range of sources: [range]
- Verification standard: [triangulation with Grade B+ sources]
## Findings
### Finding 1: [Title]
**Confidence: HIGH**
[Description with inline source references]
Sources: [Source1 (Grade A)], [Source2 (Grade B)], [Source3 (Grade A)]
### Finding 2: [Title]
**Confidence: MEDIUM**
[Description]
Note: [Source4] reports a conflicting finding. See Contradictions section.
## Contradictions and Uncertainties
[Explicit list of unresolved disagreements]
## Gaps in Evidence
[What could not be determined from available sources]
## Source Registry
[Full list of all sources with grades and verification details]
Best Practices
- Always search for disconfirming evidence, not just supporting evidence.
- Never assign HIGH confidence to a claim supported by only one source, regardless of its quality.
- Mark the date of your research prominently; findings may become outdated.
- Distinguish between "no evidence found" and "evidence of absence."
- When synthesizing, prefer systematic reviews and meta-analyses over individual studies.
- Keep the full source registry even if not all sources are cited in the final report.
References
- Bates, M. J. (1989). The Design of Browsing and Berrypicking Techniques. Online Review, 13(5), 407-424.
- Booth, A. (2006). Clear and Present Questions: Formulating Questions for Evidence-Based Practice. Library Hi Tech, 24(3), 355-368.
- Grant, M. J. & Booth, A. (2009). A Typology of Reviews. Health Information and Libraries Journal, 26(2), 91-108.
More from wentorai/research-plugins
academic-paper-summarizer
Summarize academic papers with structured extraction of key elements
41academic-translation-guide
Academic translation, post-editing, and Chinglish correction guide
30academic-writing-refiner
Checklist-driven academic English polishing and Chinglish correction
27academic-citation-manager
Manage academic citations across BibTeX, APA, MLA, and Chicago formats
26ai-writing-humanizer
Remove AI-generated patterns to produce natural, authentic academic writing
12abstract-writing-guide
Craft structured research abstracts that maximize clarity and journal acceptance
12