scraper Agent
Production-grade web scraping. Handles failures, timeouts, and rate limits.Basic Usage
Inputs
Configuration
Basic Scrape
With Custom Headers
With Retries
With Caching
Advanced Patterns
Multiple URLs with Fallback
Rate-Limited Scraping
Extract Specific Content
Scrape + Validate
Output Schema
Error Handling
The scraper agent handles:- Network timeouts: Automatic retry with exponential backoff
- 5xx errors: Retries with configurable attempts
- 4xx errors: No retry (client error)
- Redirect loops: Fails after 10 redirects
- Invalid URLs: Immediate failure
Best Practices
1. Always Set TimeoutsCommon Use Cases
Company ResearchLimitations
- JavaScript rendering: Not supported (use Puppeteer via
codeoperation) - CAPTCHA: Cannot bypass (requires manual intervention)
- Authentication: Basic only (OAuth via custom implementation)
- File downloads: Binary content not stored (use
httpoperation)

