Loading...
Loading...
Crawl websites and extract content from multiple pages via the Tavily CLI. Use this skill when the user wants to crawl a site, download documentation, extract an entire docs section, bulk-extract pages, save a site as local markdown files, or says "crawl", "get all the pages", "download the docs", "extract everything under /docs", "bulk extract", or needs content from many pages on the same domain. Supports depth/breadth control, path filtering, semantic instructions, and saving each page as a local markdown file.
npx skill4agent add tavily-ai/skills tavily-crawlcurl -fsSL https://cli.tavily.com/install.sh | bash && tvly login/docs/# Basic crawl
tvly crawl "https://docs.example.com" --json
# Save each page as a markdown file
tvly crawl "https://docs.example.com" --output-dir ./docs/
# Deeper crawl with limits
tvly crawl "https://docs.example.com" --max-depth 2 --limit 50 --json
# Filter to specific paths
tvly crawl "https://example.com" --select-paths "/api/.*,/guides/.*" --exclude-paths "/blog/.*" --json
# Semantic focus (returns relevant chunks, not full pages)
tvly crawl "https://docs.example.com" --instructions "Find authentication docs" --chunks-per-source 3 --json| Option | Description |
|---|---|
| Levels deep (1-5, default: 1) |
| Links per page (default: 20) |
| Total pages cap (default: 50) |
| Natural language guidance for semantic focus |
| Chunks per page (1-5, requires |
| |
| |
| Comma-separated regex patterns to include |
| Comma-separated regex patterns to exclude |
| Comma-separated regex for domains to include |
| Comma-separated regex for domains to exclude |
| Include external links (default: allow) |
| Include images |
| Max wait (10-150 seconds) |
| Save JSON output to file |
| Save each page as a .md file in directory |
| Structured JSON output |
--instructions--chunks-per-sourcetvly crawl "https://docs.example.com" --instructions "API authentication" --chunks-per-source 3 --json--output-dir--chunks-per-sourcetvly crawl "https://docs.example.com" --max-depth 2 --output-dir ./docs/--max-depth 1--limit 20--select-paths--limit