Scrapedo MCP Server
Web scraping for Claude Desktop, Codex, and Gemini using Scrapedo API. Simple setup with npx.
Ask AI about Scrapedo MCP Server
Powered by Claude Β· Grounded in docs
I know everything about Scrapedo MCP Server. Ask me about installation, configuration, usage, or troubleshooting.
0/500
Reviews
Documentation
Scrapedo MCP Server
Enable Claude Desktop to scrape websites using the Scrapedo API. Simple setup, powerful features.
What is this?
This is an MCP (Model Context Protocol) server that gives Claude Desktop the ability to:
- Scrape any website (with or without JavaScript)
- Take screenshots
- Use proxies from different countries
- Convert web pages to markdown
Quick Start (2 minutes)
1. Get your Scrapedo API key
Sign up at scrape.do to get your free API key.
2. Setup Claude Desktop
Run this command:
npx scrapedo-mcp-server init
That's it! The tool will automatically configure Claude Desktop for you.
3. Start using it in Claude
Just ask Claude to scrape any website:
"Can you scrape the latest news from https://example.com?"
"Take a screenshot of the Google homepage"
"Get the product prices from this e-commerce site using a US proxy"
Manual Installation Options
Install globally with npm
npm install -g scrapedo-mcp-server
scrapedo-mcp init
Install from source
git clone https://github.com/umutc/scrapedo-mcp-server.git
cd scrapedo-mcp-server
npm install
npm run build
Manual Claude Desktop configuration
Add to ~/Library/Application Support/Claude/claude_desktop_config.json:
{
"mcpServers": {
"scrapedo": {
"command": "npx",
"args": ["scrapedo-mcp-server", "start"],
"env": {
"SCRAPEDO_API_KEY": "your_api_key_here"
}
}
}
}
What Can It Do?
Basic Web Scraping
// Claude can now do this:
scrape("https://example.com")
JavaScript-Rendered Pages
// For modern SPAs and dynamic content:
scrape_with_js("https://app.example.com", {
waitSelector: ".content-loaded"
})
Screenshots
// Capture any webpage:
take_screenshot("https://example.com", {
fullPage: true
})
Residential & Mobile Proxies
// Super requests with geo-targeting:
scrape("https://example.com", {
super: true, // Switch to residential & mobile pool
geoCode: "us", // Country code (defaults to "us" if omitted)
sessionId: 42 // Optional sticky session
})
Play with Browser Actions
scrape_with_js("https://app.example.com", {
playWithBrowser: JSON.stringify([
{ action: "wait", selector: ".cookie-banner", timeout: 2000 },
{ action: "click", selector: ".accept" },
{ action: "type", selector: "#search", value: "gaming laptop" },
{ action: "press", key: "Enter" },
{ action: "wait", selector: ".results" }
])
})
Screenshots & Async Callbacks
scrape_with_js("https://example.com/product", {
screenShot: true, // or fullScreenShot / particularScreenShot
returnJSON: true, // enforced automatically when screenshot flags are set
callback: encodeURI("https://webhook.site/your-endpoint"),
customHeaders: false, // override Scrape.do default headers
forwardHeaders: true
})
Convert to Markdown
// Get clean, readable content:
scrape_to_markdown("https://blog.example.com/article")
Advanced Controls
- Toggle headers/cookies:
customHeaders,extraHeaders,forwardHeaders,setCookies,pureCookies - Switch response modes:
transparentResponse,returnJSON,showFrames,showWebsocketRequests,output: 'markdown' - Combine screenshots with scraping by setting
screenShot,fullScreenShot, orparticularScreenShot(only one at a time; automatically forcesrender+returnJSON) - Receive results later via
callbackwebhooks (Scrape.do posts JSON payloads to the provided URL)
Project Structure
scrapedo-mcp-server/
βββ src/ # TypeScript source code
β βββ index.ts # MCP server entry point
β βββ cli.ts # CLI tool for setup
β βββ tools/ # Scraping tool implementations
βββ dist/ # Compiled JavaScript (generated)
βββ package.json # Project configuration
βββ README.md # You are here
API Costs
Scrapedo charges credits per request type:
| Features | Credits Usage |
|---|---|
| Normal Request (Datacenter) | 1 |
| Datacenter + Headless Browser (JS Render) | 5 |
Residential & Mobile Request (super: true) | 10 |
| Residential & Mobile Request + Headless Browser | 25 |
Check your usage anytime by asking Claude: "Check my Scrapedo usage stats"
Common Use Cases
- Price Monitoring: Track product prices across e-commerce sites
- News Aggregation: Collect articles from multiple sources
- Data Research: Gather public information for analysis
- Content Migration: Export content from websites
- SEO Analysis: Check how pages render for search engines
- Competitive Analysis: Monitor competitor websites
FAQ
Q: Do I need to install anything?
A: Just Node.js 18+. Everything else is handled by npx.
Q: How do I update?
A: Run npx scrapedo-mcp-server@latest init to get the newest version.
Q: Is this free?
A: The MCP server is free. Scrapedo offers a free tier with limited credits.
Q: Can I use this with other MCP clients?
A: Yes! This works with any MCP-compatible client, not just Claude Desktop.
Q: How do I see debug logs?
A: Set LOG_LEVEL=DEBUG in your environment or configuration.
Q: How do I disable logging?
A: Set LOG_LEVEL=NONE to stop writing logs entirely.
Troubleshooting
| Issue | Solution |
|---|---|
| "Tool not found" | Restart Claude Desktop |
| "401 Unauthorized" | Check your API key is correct |
| "Insufficient credits" | Check usage with get_usage_stats |
| "Empty response" | The site might need JavaScript rendering |
Need Help?
- π Full API Documentation
- π¬ Report Issues
- π§ Scrapedo Support
Contributing
We welcome contributions! See CONTRIBUTING.md for guidelines.
License
MIT - See LICENSE file
Built with β€οΈ to make web scraping easy in Claude Desktop
