parallel-web-search
Web Search
Search the web for: $ARGUMENTS
Command
Choose a short, descriptive filename based on the query (e.g., ai-chip-news, react-vs-vue). Use lowercase with hyphens, no spaces.
parallel-cli search "$ARGUMENTS" -q "<keyword1>" -q "<keyword2>" --json --max-results 10 --excerpt-max-chars-total 27000 -o "$FILENAME.json"
The first argument is the objective — a natural language description of what you're looking for. It replaces multiple keyword searches with a single call for broad or complex queries. Add -q flags for specific keyword queries to supplement the objective. The -o flag saves the full results to a JSON file for follow-up questions.
Options if needed:
--after-date YYYY-MM-DDfor time-sensitive queries--include-domains domain1.com,domain2.comto limit to specific sources
Parsing results
The command outputs JSON to stdout. For each result, extract:
- title, url, publish_date
- Useful content from excerpts (skip navigation noise like menus, footers, "Skip to content")
Response format
CRITICAL: Every claim must have an inline citation. Use markdown links like Title pulling only from the JSON output. Never invent or guess URLs.
Synthesize a response that:
- Leads with the key answer/finding
- Includes specific facts, names, numbers, dates
- Cites every fact inline as Source Title — do not leave any claim uncited
- Organizes by theme if multiple topics
End with a Sources section listing every URL referenced:
Sources:
- [Source Title](https://example.com/article) (Feb 2026)
- [Another Source](https://example.com/other) (Jan 2026)
This Sources section is mandatory. Do not omit it.
After the Sources section, mention the output file path ($FILENAME.json) so the user knows it's available for follow-up questions.
If parallel-cli is not found
If the command fails with "command not found", stop immediately. Do NOT search the web yourself, do NOT use any built-in search tools, and do NOT try to answer the query from your own knowledge. Instead, tell the user:
parallel-cliis not installed- Run
/parallel-setupto install it - Then retry their search
More from michailbul/laniameda-skills
notion-sync
>
14instagram-extract
Extract text, links, and key takeaways from Instagram/Threads posts (especially carousels) and LinkedIn posts using an already-logged-in Brave/Chrome tab via OpenClaw Browser Relay. Use when the user pastes an Instagram/Threads/LinkedIn URL (or forwards screenshots) and asks something generic like “save it”, “capture this”, “summarize and store”, or “put this in the vault” — you should decide what’s worth saving, classify it into the right pillar, save it to the local KB and/or ingest prompts via the laniameda-kb skill, and also emit a compact JSON payload into a media-agent inbox file for downstream reuse.
12carousel-designer
>
11andromeda-messages
>
10deepgram-transcribe
>
10repo-kanban-pm
Install and enforce a lightweight product-management workflow inside a code repo: feature-as-kanban boards, ROADMAP status tracking, branch/PR conventions, and an optional daily OpenClaw cron PM review. Use this skill when (1) starting work in a NEW repository/project that does not yet have a ROADMAP+feature-KANBAN system, (2) you are delegated to spin up a new project and want multi-agent coordination from day one, or (3) you are asked to fix/restructure an existing repo and introduce an organized backlog/feature tracking system. Also use when adding a daily PM audit loop (cron) to keep code + docs + PRs in sync.
10