Tools / Stealth Scraper
Stealth Scraper icon

Stealth Scraper

Scrape bot-protected websites

Extract data from bot-protected websites (Cloudflare, Akamai, DataDome, PerimeterX) using residential proxies and geo-targeted IPs. Scrape a single protected page or crawl an entire bot-protected site when standard scraping fails.

2 skillsv0.01
Stealth Scrape Page

Scrape a single bot-protected web page using enhanced residential proxies, geo-targeted IPs, and extended rendering wait times. Bypasses Cloudflare, Akamai, DataDome, and similar anti-bot systems.

Returns: Scraped page content from bot-protected sites in the requested formats with metadata
Parameters
url *stringURL of the page to scrape
formatsarrayOutput formats to return (e.g. "markdown", "html", "rawHtml", "links", "screenshot")
onlyMainContentbooleanExtract only the main content, removing navbars, footers, and sidebars
includeTagsarrayCSS tags to include in extraction (e.g. ["article", "main"])
excludeTagsarrayCSS tags to exclude from extraction (e.g. ["nav", "footer"])
mobilebooleanUse a mobile user agent and viewport for rendering
waitFornumberMilliseconds to wait after page load before capturing content (default 3000)
timeoutnumberMaximum time in milliseconds to wait for the page to load (default 60000)
countrystringISO country code for geo-targeted proxy (e.g. "us", "gb", "de", "jp"). Default: "us"
languagesarrayBrowser language headers (e.g. ["en-US", "en"])
Example
Scrape a Cloudflare-protected page
curl -H "Authorization: Bearer $TOOLROUTER_API_KEY" \
  -d '{
  "tool": "stealth-scraper",
  "skill": "stealth_scrape",
  "input": {
    "url": "https://example.com/protected-page"
  }
}' \
  https://api.toolrouter.com/v1/tools/call
Stealth Crawl Site

Recursively crawl a bot-protected website using enhanced proxies on every page. Bypasses anti-bot systems across the entire crawl, with geo-targeted IPs and extended rendering.

Returns: Array of scraped pages from bot-protected sites with content, metadata, and crawl status
Parameters
url *stringStarting URL for the crawl
limitnumberMaximum number of pages to crawl (default 50)
maxDepthnumberMaximum link-following depth from the starting URL
includePathsarrayURL path patterns to include (e.g. ["/blog/*", "/docs/*"])
excludePathsarrayURL path patterns to exclude (e.g. ["/admin/*", "/api/*"])
allowSubdomainsbooleanWhether to follow links to subdomains of the starting URL
allowExternalLinksbooleanWhether to follow links to external domains
countrystringISO country code for geo-targeted proxy (e.g. "us", "gb", "de", "jp"). Default: "us"
languagesarrayBrowser language headers (e.g. ["en-US", "en"])
Example
Crawl a protected blog section
curl -H "Authorization: Bearer $TOOLROUTER_API_KEY" \
  -d '{
  "tool": "stealth-scraper",
  "skill": "stealth_crawl",
  "input": {
    "url": "https://example.com/blog",
    "limit": 20
  }
}' \
  https://api.toolrouter.com/v1/tools/call
Loading reviews...
Loading activity...
v0.012026-03-23
  • Initial release with stealth_scrape and stealth_crawl skills

Quick Start

MCP (Claude Code)
claude mcp add --transport stdio \
  --env TOOLROUTER_API_KEY=YOUR_API_KEY \
  toolrouter -- npx -y toolrouter-mcp
REST API
curl -H "Authorization: Bearer $TOOLROUTER_API_KEY" \
  -d '{"tool":"stealth-scraper","skill":"stealth_scrape","input":{}}' \
  https://api.toolrouter.com/v1/tools/call

Frequently Asked Questions

When should I use this instead of a normal scraper?

Use `stealth_scrape` when a site throws 403s, 429s, CAPTCHA pages, or bot protection from Cloudflare, Akamai, DataDome, or PerimeterX.

Can it scrape just one page or an entire site?

`stealth_scrape` handles a single protected page, while `stealth_crawl` recursively crawls a whole site with a page limit and depth cap.

Can I match the browser to the site region or language?

Yes. Set `country` for geo-targeted proxies and `languages` for locale headers. You can also use `mobile`, `waitFor`, and `onlyMainContent` when the site needs it.

What output formats can I get?

`stealth_scrape` can return formats like markdown, HTML, raw HTML, links, and screenshots, along with the extracted page metadata.