How much bandwidth does web scraping typically use?

Bandwidth usage varies dramatically based on website type and scraping approach.

Typical page sizes by content type:

  • Text-heavy sites (news, blogs): 100-500 KB
  • E-commerce product pages: 1-3 MB (with images)
  • Social media feeds: 2-5 MB (images, videos, ads)
  • Media-heavy sites (galleries): 5-20 MB per page

Scraping approach impact:

  • Headless browser (full page load): 100% of page size
  • Headless with resource blocking: 10-30% of page size
  • HTTP requests only (no browser): 5-15% of page size
  • API endpoint: 1-5 KB per request (most efficient)

Real-world examples:

  • Scraping 10,000 e-commerce products with headless browser: ~20 GB
  • Same scrape with images blocked: ~3 GB
  • Same scrape using HTTP only: ~500 MB
  • Same scrape using product API: ~50 MB

Factors that increase bandwidth:

  • Loading unnecessary assets (images, fonts, analytics)
  • JavaScript-heavy single-page applications
  • Auto-playing videos or animations
  • High-resolution images
  • Multiple failed requests and retries

Optimization opportunities:

Most scrapers waste 80-90% of bandwidth on resources they don't need. A bandwidth calculator helps identify exactly which resources you're loading and which can be safely blocked.

Monthly estimates:

  • Small project (1,000 pages/day): 60-200 GB/month
  • Medium project (10,000 pages/day): 600 GB - 2 TB/month
  • Large project (100,000 pages/day): 6-20 TB/month

These ranges assume moderate optimization. Without optimization, costs can be 5-10x higher.

Related Questions