Introduction

The ScraperAPI Crawler is ideal for scraping websites, where the data you want to extract spans across multiple linked pages. It’s great for extracting data from product listings, paginated search results, real estate catalogs - basically any structure, where one page leads to dozens or hundreds more. It takes care of crawling, scraping, retries, and delivers the results back to you (supports webhook callbacks too).

What the Crawler Does

  • Discovers and scrapes new pages based on how you configure it.

  • Skips duplicates to avoid infinite loops.

  • Handles failed requests gracefully.

  • Stops when your credit budget or depth limit is hit.

  • Streams page results during the crawl and sends a full summary at the end.

Whether you're crawling 10 pages or 10,000, it runs the job from start to finish and saves each page result in real time (or sends it over to your webhook).

Last updated

Was this helpful?