Overview
WhizoAI’s lightweight engine provides blazing-fast scraping for static websites that don’t require JavaScript rendering. Perfect for content-heavy sites, APIs, and bulk data collection.Lightweight vs Browser Engines
| Feature | Lightweight | Playwright/Puppeteer |
|---|---|---|
| Speed | ⚡ 5-10x faster | Standard |
| Cost | 1 credit | 2 credits (with JS) |
| JavaScript | ❌ No | ✅ Yes |
| Best For | Static HTML, APIs | SPAs, Dynamic content |
| Memory | Low | High |
| Concurrent Requests | Very High | Limited |
Basic Usage
Performance Optimization
Concurrent Processing
Process multiple URLs simultaneously:HTTP/2 Support
Enable HTTP/2 for better performance:Compression
Reduce bandwidth with compression:Selective Content Loading
Block Unnecessary Resources
Target Specific Elements
Only extract what you need:Caching Strategy
Enable Result Caching
Cache Control Headers
Respect site’s cache headers:Timeout Management
Optimize Timeouts
Parallel Timeout Strategy
Format Optimization
Choose Efficient Formats
Batch Processing Best Practices
Chunked Processing
Process large batches in chunks:Rate Limit Management
Respect rate limits:Performance Benchmarks
Speed Comparison
| Pages | Lightweight | Playwright | Time Saved |
|---|---|---|---|
| 10 | 5 seconds | 30 seconds | 83% |
| 100 | 45 seconds | 5 minutes | 85% |
| 1000 | 7 minutes | 50 minutes | 86% |
Cost Comparison
| Feature | Lightweight | With JavaScript |
|---|---|---|
| Base scraping | 1 credit | 2 credits |
| 1000 pages | 1000 credits | 2000 credits |
| Savings | - | 50% |
When to Use Lightweight Engine
Static Websites
Static Websites
News sites, blogs, documentation sites, government portals
API Scraping
API Scraping
REST APIs, XML feeds, JSON endpoints
Bulk Data Collection
Bulk Data Collection
Large-scale scraping where speed matters
Server-Side Rendered Pages
Server-Side Rendered Pages
Traditional server-rendered HTML (PHP, Ruby, Python backends)
When NOT to Use Lightweight
Monitoring Performance
Common Optimizations
Reduce Payload
Block images, fonts, and ads to load only essential content
Parallel Processing
Use concurrent requests to maximize throughput
Smart Caching
Cache frequently accessed pages to save credits
Efficient Formats
Choose text/markdown over HTML for smaller payloads
Related Resources
Batch Processing
Process thousands of pages efficiently
Browser Automation
When you need JavaScript rendering
API Reference
Full scraping API documentation