The ScrapeOps n8n node is a powerful integration that brings ScrapeOps' web scraping capabilities directly into your n8n workflows. This node allows you to leverage ScrapeOps' proxy aggregation, HTML parsing, and structured data extraction services without writing complex code.
ScrapeOps provides a suite of tools to simplify web scraping operations:
- π Proxy API: Access high-quality proxies through a single endpoint for reliable web scraping
- π Parser API: Extract structured data from HTML content of popular websites
- πΎ Data API: Retrieve and query data from ScrapeOps datasets
- Smart proxy rotation across multiple providers
- JavaScript rendering support for dynamic sites
- Anti-bot bypass capabilities (Cloudflare, DataDome, etc.)
- Geo-targeting options
- Mobile and residential proxy support
- Session management for multi-step scraping
- Extract structured data from popular websites without maintaining your own parsers
- Supported Sites: Amazon, eBay, Walmart, Indeed, Redfin
- Page Types: Product pages, search results, reviews, categories
- Returns clean, structured JSON data
- Direct access to structured data endpoints
- Amazon Product API: Get product details by ASIN or URL
- Amazon Search API: Search products and get structured results
- No HTML scraping required - get data in a single request
If you're using n8n Cloud, you can easily install the ScrapeOps node directly from the UI:
- Sign in to n8n, open the editor, and click + in the top right to open the Nodes panel
- Search for "ScrapeOps" node using the search bar. Look for the version marked by a badge β. Then, select install.
- The ScrapeOps node will be installed and appear in your node palette automatically.
For self-hosted n8n instances:
- Open your n8n instance
- Navigate to Settings β Community Nodes
- Click Install a community node
- Enter the package name:
@scrapeops/n8n-nodes-scrapeops
- Click Install
- Restart your n8n instance when prompted
# Navigate to your n8n installation directory
cd ~/.n8n
# Install the ScrapeOps node
npm install @scrapeops/n8n-nodes-scrapeops
# Restart n8n
n8n start
For Docker users, add the node to your docker-compose.yml
:
version: '3.8'
services:
n8n:
image: n8nio/n8n
environment:
- N8N_COMMUNITY_NODES_ENABLED=true
- NODE_FUNCTION_ALLOW_EXTERNAL=n8n-nodes-scrapeops
volumes:
- ~/.n8n:/home/node/.n8n
Then install the node:
docker exec -it <container_name> npm install @scrapeops/n8n-nodes-scrapeops
docker restart <container_name>
To use the ScrapeOps node, you'll need a ScrapeOps API key which you can get by signing up for a free account here.
Steps to get your API key:
- Sign up for a free account at ScrapeOps
- Verify your email address (required to activate your API key)
- Visit your dashboard at ScrapeOps Dashboard
- Copy your API key from the dashboard
β οΈ Important: You must confirm your email address to activate your API key. Check your inbox for a verification email from ScrapeOps.
- In n8n, go to Credentials β Add Credential.
- Search for "ScrapeOps API" and enter your API key.
- Save and test the credentials.
- Create a new workflow in n8n.
- Click Add to Workflow "ScrapeOps" node from the palette.
- Select an API (Proxy, Parser, or Data) and configure parameters.
You can monitor your usage in the ScrapeOps Dashboard.
Route GET/POST requests through proxies to scrape blocked sites.
Basic Parameters:
- URL: Target URL to scrape (required)
- Method: GET or POST (default: GET)
- Return Type: Default (raw response) or JSON
Advanced Options:
Option | Type | Description | Default | Example Values |
---|---|---|---|---|
Follow Redirects | Boolean | Follow HTTP redirects | true | true , false |
Keep Headers | Boolean | Use your custom headers | false | true , false |
Initial Status Code | Boolean | Return initial status code | false | true , false |
Final Status Code | Boolean | Return final status code | false | true , false |
Optimize Request | Boolean | Auto-optimize settings | false | true , false |
Max Request Cost | Number | Max credits to use (with optimize) | 0 | 10 , 50 , 100 |
Render JavaScript | Boolean | Enable headless browser | false | true , false |
Wait Time | Number | Wait before capture (ms) | 0 | 3000 , 5000 |
Wait For | String | CSS selector to wait for | - | .product-title , #content |
Scroll | Number | Scroll pixels before capture | 0 | 1000 , 2000 |
Screenshot | Boolean | Return base64 screenshot | false | true , false |
Device Type | String | Device emulation | desktop | desktop , mobile |
Premium Proxies | String | Premium level | level_1 | level_1 , level_2 |
Residential Proxies | Boolean | Use residential IPs | false | true , false |
Mobile Proxies | Boolean | Use mobile IPs | false | true , false |
Session Number | Number | Sticky session ID | 0 | 12345 , 67890 |
Country | String | Geo-targeting country | - | us , gb , de , fr , ca , au , jp , in |
Bypass | String | Anti-bot bypass level | - | cloudflare_level_1 , cloudflare_level_2 , cloudflare_level_3 , datadome , perimeterx , incapsula , generic_level_1 to generic_level_4 |
Example Configuration:
API Type: Proxy API
URL: https://example.com
Method: GET
Render JavaScript: true
Wait Time: 3000
Country: us
Parse HTML into structured JSON for supported domains.
Parameters:
- Domain: Amazon, eBay, Walmart, Indeed, Redfin
- Page Type: Varies by domain (e.g., Product, Search for Amazon)
- URL: Page URL (required)
- HTML Content: Raw HTML to parse (required)
Example Configuration:
API Type: Parser API
Domain: Amazon
Page Type: Product Page
URL: https://www.amazon.com/dp/B08N5WRWNW
HTML Content: {{ $node["Proxy_API"].json.body }}
Access pre-scraped datasets, focused on Amazon.
Parameters:
- Domain: Amazon (more coming soon)
- Amazon API Type: Product or Search
- Input Type: ASIN/URL for Product; Query/URL for Search
Example Configuration:
API Type: Data API
Domain: Amazon
Amazon API Type: Product API
Input Type: ASIN
ASIN: B08N5WRWNW
- Track product prices across multiple e-commerce sites
- Set up alerts for price drops
- Generate competitive pricing reports
- Collect product reviews and ratings
- Monitor competitor inventory
- Track market trends
- Extract business information from directories
- Collect job postings from career sites
- Find real estate listings
- Collect news articles
- Monitor social media mentions
- Aggregate product descriptions
Begin with basic requests and add complexity:
1. Basic GET request
2. Add JavaScript rendering if needed
3. Add wait conditions
4. Enable anti-bot bypass if blocked
- Use
optimize_request: true
for automatic optimization - Set
max_request_cost
to control spending - Only enable features you need
For JavaScript-heavy sites:
Render JavaScript: true
Wait Time: 3000
Wait For: .main-content
Scroll: 1000
For multi-step scraping:
Session Number: 12345
// Use same number across requests
Problem: ScrapeOps node doesn't show up after installation
Solution:
- Ensure n8n was restarted after installation
- Check that community nodes are enabled
- Verify the installation with:
npm list @scrapeops/n8n-nodes-scrapeops
Problem: "Invalid API Key" error
Solution:
- Verify API key is copied correctly (no extra spaces)
- Check if API key is active in ScrapeOps dashboard
- Ensure you're using the correct credential in the node
Problem: Requests timing out
Solution:
- Check your firewall settings
- Verify n8n can make external HTTP requests
- Test with a simple URL first (like httpbin.org)
- ScrapeOps Documentation - Complete API documentation
- ScrapeOps n8n Node Documentation - Complete ScrapeOps n8n Node documentation
- n8n Community Nodes Documentation - n8n integration guide
- ScrapeOps Dashboard - Monitor usage and performance
- n8n Community Forum - Get help from the community
- ScrapeOps Support: [email protected]
- n8n Community Forum: community.n8n.io
- GitHub Issues: Report bugs or request features
Ready to automate your web scraping? Get started with the ScrapeOps n8n integration today!