search1api

SKILL.md

Search1API CLI

Web search and content retrieval via the s1 command-line tool (search1api-cli).

Prerequisites

Before using any command, check if s1 is available. If not, guide the user to install it:

npm install -g search1api-cli

An API key is also required. Get one at https://search1api.com, then configure:

s1 config set-key <your-api-key>

Or set the environment variable SEARCH1API_KEY.

If a command fails with "command not found" or an auth error, remind the user to complete these setup steps before retrying.

When to use

User intent Command
Shares a URL / link → read and summarize s1 crawl <url>
Wants to search the web s1 search "<query>"
Wants news s1 news "<query>"
Wants to explore a site's links s1 sitemap <url>
Wants trending topics s1 trending <service>
Wants deep thinking on a problem s1 reasoning "<content>"
Wants to check remaining credits s1 balance

Dynamic tuning

Adapt parameters to user intent — don't just use defaults:

  • Quick lookup ("search for X", "what is X") → -n 5, no crawl
  • Deep research ("research X thoroughly", "comprehensive analysis") → -n 15, then crawl top 3–5 results with separate s1 crawl calls
  • User specifies a number ("find 10 articles") → match it with -n
  • Recency signals ("latest", "recent", "this week") → -t day or -t month
  • Domain-specific ("search on Reddit", "find GitHub repos") → -s reddit, -s github, etc.
  • Site-scoped ("only from arxiv.org") → --include arxiv.org
  • Chinese queries → consider -s baidu for better results

Commands

search

s1 search "<query>" [options]
Option Description Default
-n, --max-results <N> Number of results (1–50) 10
-s, --service <engine> Search engine google
-c, --crawl <N> Crawl N results for full content 0
--include <sites...> Only include these sites
--exclude <sites...> Exclude these sites
-t, --time <range> day, month, year
--json Raw JSON output

Search engines: google, bing, duckduckgo, yahoo, x, reddit, github, youtube, arxiv, wechat, bilibili, imdb, wikipedia

news

s1 news "<query>" [options]

Same options as search. News services: google, bing, duckduckgo, yahoo, hackernews. Default service: bing.

When user asks for breaking/latest news, always add -t day.

crawl

s1 crawl <url>

Extracts clean content from a URL. Use this whenever the user shares a link.

sitemap

s1 sitemap <url>

Returns all discovered links on a URL/domain.

reasoning

s1 reasoning "<content>"
# or
s1 reason "<content>"

Deep thinking powered by DeepSeek R1. Use for complex analytical questions.

trending

s1 trending <service> [-n <N>]

Services: github, hackernews.

balance

s1 balance

Shows remaining API credits.

Workflows

Deep research

  1. s1 search "<topic>" -n 15 → get broad results
  2. s1 crawl <url> → crawl the top 3–5 most relevant URLs from results
  3. Synthesize all gathered content into a coherent answer with source citations

URL summarization

  1. s1 crawl <url> → get the page content
  2. Summarize or answer questions based on the content

Trending deep dive

  1. s1 trending github -n 10 → discover hot topics
  2. s1 search "<interesting topic>" -t day → search for details
  3. s1 crawl <url> → read full article if needed

Output handling

  • By default, commands produce human-readable formatted output
  • Add --json to any command for raw JSON (useful for programmatic processing)
  • After retrieving results, always summarize and synthesize the information for the user — don't just dump raw output

References

Weekly Installs
2
GitHub Stars
1
First Seen
5 days ago
Installed on
cline2
github-copilot2
codex2
kimi-cli2
gemini-cli2
cursor2