Welcome to SpiderIQ Guides
This section provides comprehensive guides and tutorials to help you get the most out of SpiderIQ’s web scraping and Google Maps data extraction capabilities.What is SpiderIQ?
SpiderIQ is a high-performance API service that provides four specialized capabilities:SpiderSite
Website ScrapingExtract content from any website using the Crawl4AI library with optional AI-powered data extraction.
- Full-page markdown conversion
- AI-powered content extraction
- Screenshot capture
- Metadata extraction
SpiderMaps
Google Maps ScrapingExtract business information from Google Maps using Playwright browser automation.
- Business details (name, address, phone)
- Reviews and ratings
- Business hours
- Categories and photos
- Campaign System (v2.14.0): Multi-location orchestration
SpiderVerify
Email VerificationVerify email addresses at the SMTP level without sending actual emails.
- Deliverability checking
- Disposable email detection
- Role account identification
- Quality scoring (0-100)
SpiderPeople
Decision Maker Discovery (v2.17.0)Find the right people behind companies using ICP-based search.
- Natural language search by role + location
- Profile lookup by LinkedIn URL
- AI research reports
- Experience & education data
Quick Links
Getting Started
Submit your first job in 5 minutes
Authentication
Learn about API authentication
API Reference
Complete API documentation
Available Guides
Scraping Websites
Complete guide to website scraping with SpiderSite
Google Maps
Extract business data from Google Maps
Email Verification
Verify emails via SMTP without sending
People Research
Research LinkedIn profiles with AI insights
v2.18.0: FuzzIQ Deduplication
FuzzIQ Per-Client Deduplication
NEW Automatic record deduplication across all job types with per-client data isolation.
- Per-Record Unique Flag: Each record marked with
fuzziq_unique: true/false - Client Settings: Configure
fuzziq_enabledandfuzziq_unique_onlyper client - Response Filtering: Use
fuzziq_unique_only: trueto return only new records - Isolated Schemas: Separate PostgreSQL schemas per client for complete data isolation
v2.15.0: Orchestrated Campaigns
Orchestrated Campaigns
Chain SpiderMaps + SpiderSite + SpiderVerify in a single workflow
Xano Integration
Build lead gen systems with Xano no-code backend
n8n Integration
Automate campaigns with n8n workflow automation
Common Use Cases
Content Aggregation
Extract articles, blog posts, and documentation from multiple sources for content analysis or aggregation platforms. Example: News monitoring, competitor content analysis, research aggregationE-commerce Data
Scrape product information, prices, and reviews from e-commerce sites for price monitoring or market research. Example: Price comparison tools, inventory monitoring, product catalog buildingLocal Business Research
Extract business information from Google Maps for lead generation, market research, or directory creation. Example: B2B prospecting, competitive analysis, local SEO researchReal Estate & Property Data
Gather property listings, prices, and details for real estate analysis and market trends. Example: Property aggregators, market analysis tools, investment researchJob Board Aggregation
Collect job postings from multiple sources to create comprehensive job search platforms. Example: Job aggregators, salary analysis, hiring trend researchHow SpiderIQ Works
Processing Flow
- Submit - Client submits a job via API
- Queue - Job is queued for processing
- Process - Available worker picks up and processes the job
- Store - Results are saved (screenshots to Cloudflare R2, data to Database)
- Retrieve - Client polls for results and receives data
Architecture
SpiderIQ is built on a scalable, distributed architecture:- API Gateway - FastAPI-based REST API
- Message Queue - Job distribution system
- Workers - Distributed scraping workers (Docker containers)
- Database - Database for job metadata and results
- Cache - Redis for performance optimization
- CDN Storage - Cloudflare R2 for screenshots
Worker Types
- SpiderSite Workers - 70 workers for website scraping
- SpiderMaps Workers - 42 workers for Google Maps scraping
- SpiderVerify Workers - 10 workers for email verification
- SpiderPeople Workers - 1 worker for LinkedIn research
Performance & Limits
Rate Limits
Standard Rate Limit: 100 requests per minute per clientBurst allowance of 20 requests for occasional spikes. Contact us for higher limits.
Processing Times
| Job Type | Average Time | Range |
|---|---|---|
| SpiderSite (simple page) | 5-15s | 3-30s |
| SpiderSite (with AI) | 10-25s | 5-45s |
| SpiderMaps | 3-8s | 2-15s |
| SpiderVerify (single) | 2-5s | 1-10s |
| SpiderVerify (bulk 100) | 30-60s | 20-120s |
| SpiderPeople (profile) | 5-10s | 3-15s |
| SpiderPeople (search) | 5-15s | 3-20s |
| SpiderPeople (research) | 15-30s | 10-45s |
Queue Capacity
- Normal load: < 20 jobs queued
- Moderate load: 20-50 jobs queued
- High load: > 50 jobs queued
Best Practices
Need Help?
API Reference
Complete API documentation with all endpoints
Support
Contact our support team
System Status
Check API health and queue stats
Get API Access
Request API credentials
Next Steps
1
Get Credentials
Contact admin@di-atomic.com to get your API credentials
2
Read the Quickstart
Follow our 5-minute quickstart guide to submit your first job
3
Explore Guides
Learn about website scraping and explore the API reference
4
Build Your Integration
Use the API reference to build your integration
