tavily-tools

SKILL.md

[H1][TAVILY-TOOLS]

Dictum: Command-specific arguments enforce correct invocation.

Execute Tavily AI web operations through unified Python CLI.

[IMPORTANT] search requires --query; extract requires --urls; crawl/map require --url. 1Password injects API key automatically.

# Search
uv run .claude/skills/tavily-tools/scripts/tavily.py search --query "Vite 7 new features"
uv run .claude/skills/tavily-tools/scripts/tavily.py search --query "React 19" --topic news
uv run .claude/skills/tavily-tools/scripts/tavily.py search --query "Effect-TS" --search-depth advanced --max-results 20

# Extract content from URLs
uv run .claude/skills/tavily-tools/scripts/tavily.py extract --urls "https://example.com"
uv run .claude/skills/tavily-tools/scripts/tavily.py extract --urls "https://a.com,https://b.com" --format text

# Crawl website
uv run .claude/skills/tavily-tools/scripts/tavily.py crawl --url "https://docs.effect.website"
uv run .claude/skills/tavily-tools/scripts/tavily.py crawl --url "https://nx.dev" --max-depth 3 --max-breadth 50

# Map site structure
uv run .claude/skills/tavily-tools/scripts/tavily.py map --url "https://nx.dev"
uv run .claude/skills/tavily-tools/scripts/tavily.py map --url "https://effect.website" --max-depth 2 --limit 200

[1][OUTPUT]

Commands return: {"status": "success|error", ...}.

[INDEX] [CMD] [RESPONSE]
[1] search {query, results[], images[], answer}
[2] extract {urls[], results[], failed[]}
[3] crawl {base_url, results[], urls_crawled}
[4] map {base_url, urls[], total_mapped}
Weekly Installs
2
Installed on
claude-code2
windsurf1
opencode1
codex1
antigravity1
gemini-cli1