Crawl websites

Create a crawl job, wait for completion, then iterate pages. Scope with patterns and limit with depth/link_limit.

Note: Install the SDKs if you haven’t already: https://docs.supacrawler.com/api/install

Quickstart

curl https://api.supacrawler.com/api/v1/crawl \
  -H "Authorization: Bearer YOUR_API_KEY" \
  -H "Content-Type: application/json" \
  -d '{
    "url": "https://example.com/docs",
    "type": "crawl",
    "depth": 2,
    "link_limit": 20,
    "patterns": ["/docs/*"],
    "format": "markdown"
  }'

Tips:

  • Narrow the scope with patterns to avoid crawling marketing/blog
  • Raise depth gradually; start small to verify targets
  • Consider render_js for SPA docs if needed

Links: Jobs API

Was this page helpful?