notion-meeting-intelligence
Meeting Intelligence
Prep meetings by pulling Notion context, tailoring agendas/pre-reads, and enriching with Codex research.
Quick start
- Confirm meeting goal, attendees, date/time, and decisions needed.
- Gather context: search with
Notion:notion-search, then fetch withNotion:notion-fetch(prior notes, specs, OKRs, decisions). - Pick the right template via
reference/template-selection-guide.md(status, decision, planning, retro, 1:1, brainstorming). - Draft agenda/pre-read in Notion with
Notion:notion-create-pages, embedding source links and owner/timeboxes. - Enrich with Codex research (industry insights, benchmarks, risks) and update the page with
Notion:notion-update-pageas plans change.
Workflow
0) If any MCP call fails because Notion MCP is not connected, pause and set it up:
- Add the Notion MCP:
codex mcp add notion --url https://mcp.notion.com/mcp
- Enable remote MCP client:
- Set
[features].rmcp_client = trueinconfig.tomlor runcodex --enable rmcp_client
- Set
- Log in with OAuth:
codex mcp login notion
After successful login, the user will have to restart codex. You should finish your answer and tell them so when they try again they can continue with Step 1.
1) Gather inputs
- Ask for objective, desired outcomes/decisions, attendees, duration, date/time, and prior materials.
- Search Notion for relevant docs, past notes, specs, and action items (
Notion:notion-search), then fetch key pages (Notion:notion-fetch). - Capture blockers/risks and open questions up front.
2) Choose format
- Status/update → status template.
- Decision/approval → decision template.
- Planning (sprint/project) → planning template.
- Retro/feedback → retrospective template.
- 1:1 → one-on-one template.
- Ideation → brainstorming template.
- Use
reference/template-selection-guide.mdto confirm.
3) Build the agenda/pre-read
- Start from the chosen template in
reference/and adapt sections (context, goals, agenda, owner/time per item, decisions, risks, prep asks). - Include links to pulled Notion pages and any required pre-reading.
- Assign owners for each agenda item; call out timeboxes and expected outputs.
4) Enrich with research
- Add concise Codex research where helpful: market/industry facts, benchmarks, risks, best practices.
- Keep claims cited with source links; separate fact from opinion.
5) Finalize and share
- Add next steps and owners for follow-ups.
- If tasks arise, create/link tasks in the relevant Notion database.
- Update the page via
Notion:notion-update-pagewhen details change; keep a brief changelog if multiple edits.
References and examples
reference/— template picker and meeting templates (e.g.,template-selection-guide.md,status-update-template.md,decision-meeting-template.md,sprint-planning-template.md,one-on-one-template.md,retrospective-template.md,brainstorming-template.md).examples/— end-to-end meeting preps (e.g.,executive-review.md,project-decision.md,sprint-planning.md,customer-meeting.md).
More from cuba6112/skillfactory
ollama-rag
Build RAG systems with Ollama local + cloud models. Latest cloud models include DeepSeek-V3.2 (GPT-5 level), Qwen3-Coder-480B (1M context), MiniMax-M2. Use for document Q&A, knowledge bases, and agentic RAG. Covers LangChain, LlamaIndex, ChromaDB, and embedding models.
17unsloth-sft
Supervised fine-tuning using SFTTrainer, instruction formatting, and multi-turn dataset preparation with triggers like sft, instruction tuning, chat templates, sharegpt, alpaca, conversation_extension, and SFTTrainer.
6torchaudio
Audio signal processing library for PyTorch. Covers feature extraction (spectrograms, mel-scale), waveform manipulation, and GPU-accelerated data augmentation techniques. (torchaudio, melscale, spectrogram, pitchshift, specaugment, waveform, resample)
5pytorch-onnx
Exporting PyTorch models to ONNX format for cross-platform deployment. Includes handling dynamic axes, graph optimization in ONNX Runtime, and INT8 model quantization. (onnx, onnxruntime, torch.onnx.export, dynamic_axes, constant-folding, edge-deployment)
5unsloth-lora
Configuring and optimizing 16-bit Low-Rank Adaptation (LoRA) and Rank-Stabilized LoRA (rsLoRA) for efficient LLM fine-tuning using triggers like lora, qlora, rslora, rank selection, lora_alpha, lora_dropout, and target_modules.
4torchvision
Computer vision library for PyTorch featuring pretrained models, advanced image transforms (v2), and utilities for handling complex data types like bounding boxes and masks. (torchvision, transforms, tvtensor, resnet, cutmix, mixup, pretrained models, vision transforms)
3