reviewing-ai-papers
Reviewing AI Papers
When users request analysis of AI/ML technical content (papers, articles, blog posts), extract actionable insights filtered through an enterprise AI engineering lens and store valuable discoveries to memory for cross-session recall.
Contextual Priorities
Technical Architecture:
- RAG systems (semantic/lexical search, hybrid retrieval)
- Vector database optimization and embedding strategies
- Model fine-tuning for specialized scientific domains
- Knowledge distillation for secure on-premise deployment
Implementation & Operations:
- Prompt engineering and in-context learning techniques
- Security and IP protection in AI systems
- Scientific accuracy and hallucination mitigation
- AWS integration (Bedrock/SageMaker)
Enterprise & Adoption:
- Enterprise deployment in regulated environments
- Building trust with scientific/legal stakeholders
- Internal customer success strategies
- Build vs. buy decision frameworks
Analytical Standards
- Maintain objectivity: Extract factual insights without amplifying source hype
- Challenge novelty claims: Identify what practitioners already use as baselines. Distinguish "applies existing techniques" from "genuinely new methods"
- Separate rigor from novelty: Well-executed study of standard techniques ≠ methodological breakthrough
- Confidence transparency: Distinguish established facts, emerging trends, speculative claims
- Contextual filtering: Prioritize insights mapping to current challenges
Analysis Structure
For Substantive Content
Article Assessment (2-3 sentences)
- Core topic and primary claims
- Credibility: author expertise, evidence quality, methodology rigor
Prioritized Insights
- High Priority: Direct applications to active projects
- Medium Priority: Adjacent technologies worth monitoring
- Low Priority: Interesting but not immediately actionable
Technical Evaluation
- Distinguish novel methods from standard practice presented as innovation
- Flag implementation challenges, risks, resource requirements
- Note contradictions with established best practices
Actionable Recommendations
- Research deeper: Specific areas requiring investigation
- Evaluate for implementation: Techniques worth prototyping
- Share with teams: Which teams benefit from this content
- Monitor trends: Emerging areas to track
Immediate Applications Map insights to current projects. Identify quick wins or POC opportunities.
For Thin Content
- State limitations upfront
- Extract marginal insights if any
- Recommend alternatives if topic matters
- Keep brief
Memory Integration
Automatic storage triggers:
- High-priority insights (directly applicable)
- Novel techniques worth prototyping
- Pattern recognitions across papers
- Contradictions to established practice
Storage format:
remember(
"[Source: {title or url}] {condensed insight}",
"world",
tags=["paper-insight", "{domain}", "{technique}"],
conf=0.85 # higher for strong evidence
)
Compression rule:
- Full analysis → conversation (what user sees)
- Condensed insight → memory (searchable nugget with attribution)
- Store the actionable kernel, not the whole analysis
Example:
Analysis says: "Hybrid retrieval (BM25 + dense) shows 23% improvement over pure semantic search for scientific queries. Two-stage approach..."
Store as: "[Source: arxiv.org/abs/2401.xxxxx] Hybrid BM25+dense retrieval: 23% lift over semantic-only for scientific corpora. Requires 10K+ domain examples for fine-tuning benefit."
Tags: ["paper-insight", "rag", "hybrid-retrieval", "scientific-domain"]
Output Standards
- Conciseness: Actionable insights, not content restatement
- Precision: Distinguish demonstrates/suggests/claims/speculates
- Relevance: Connect to focus areas or state no connection
- Adaptive depth: Match length to content value
Constraints
- No hype amplification
- No timelines unless requested
- No speculation beyond article
- Note contradictions explicitly
- State limitations on thin content
More from oaustegard/claude-skills
developing-preact
Specialized Preact development skill for standards-based web applications with native-first architecture and minimal dependency footprint. Use when building Preact projects, particularly those involving data visualization, interactive applications, single-page apps with HTM syntax, Web Components integration, CSV/JSON data parsing, WebGL shader visualizations, or zero-build solutions with vendored ESM imports.
104exploring-codebases
>-
62mapping-codebases
Generate navigable code maps for unfamiliar codebases. Extracts exports/imports via AST (tree-sitter) to create _MAP.md files per directory showing classes, functions, methods with signatures and line numbers. Use when exploring repositories, understanding project structure, analyzing unfamiliar code, or before modifications. Triggers on "map this codebase", "explore repo", "understand structure", "what does this project contain", or when starting work on an unfamiliar repository.
48accessing-github-repos
GitHub repository access in containerized environments using REST API and credential detection. Use when git clone fails, or when accessing private repos/writing files via API.
43remembering
Advanced memory operations reference. Basic patterns (profile loading, simple recall/remember) are in project instructions. Consult this skill for background writes, memory versioning, complex queries, edge cases, session scoping, retention management, type-safe results, proactive memory hints, GitHub access detection, autonomous curation, episodic scoring, and decision traces.
41asking-questions
Guidance for asking clarifying questions when user requests are ambiguous, have multiple valid approaches, or require critical decisions. Use when implementation choices exist that could significantly affect outcomes.
40