MCP Server
Use Deepcrawl directly from Claude, ChatGPT, Cursor, and other AI assistants via the Model Context Protocol.
Deepcrawl provides an MCP (Model Context Protocol) server that lets AI assistants like Claude, ChatGPT, Cursor, and others directly call Deepcrawl's web scraping and content extraction tools without any code.
Quick Start
Add Deepcrawl to your MCP configuration:
{
"mcpServers": {
"deepcrawl": {
"command": "npx",
"args": ["mcp-remote", "https://deepcrawl-mcp.shinzero.workers.dev/mcp"]
}
}
}The MCP server is publicly accessible - no additional authentication required. It uses your DeepCrawl API key automatically for all requests.
Available Tools
read_url
Read content from any URL and get structured data including HTML, markdown, and metadata.
Parameters:
url(required) - The URL to readmarkdown(optional) - Return markdown contentrawHtml(optional) - Return raw HTML
Example:
read_url from https://example.comget_markdown
Get clean markdown content from any URL. Ideal for extracting readable content from articles, documentation, or any text-heavy page.
Parameters:
url(required) - The URL to extract markdown fromcleanHTML(optional) - Apply additional HTML cleaning
Example:
Get markdown from https://example.comextract_links
Extract all links from a webpage. Returns internal links, external links, and media URLs in a structured format.
Parameters:
url(required) - The URL to extract links fromincludeExternal(optional) - Include external linksincludeMedia(optional) - Include media URLs (images, videos)tree(optional) - Return links as a hierarchical tree
Example:
Extract all links from https://example.com as a treelist_logs
View your recent API activity logs. See what requests have been made, their status, and timestamps.
Parameters:
limit(optional) - Number of logs to return (default: 10)startDate(optional) - Filter by start date (ISO format)endDate(optional) - Filter by end date (ISO format)
Example:
Show my last 20 API logsUse Cases
Research Agent
Use Deepcrawl as part of an AI research workflow:
- Map a site - Use
extract_linksto understand a website's structure - Get content - Use
get_markdownfor quick summaries - Deep dive - Use
read_urlwhen you need full metadata or HTML
Content Extraction
Quickly extract content from web pages:
Get the main content from https://docs.example.com/api-guide as markdownLink Analysis
Analyze a website's link structure:
Extract all internal links from https://example.com as a treeConfiguration
The MCP server is hosted at:
https://deepcrawl-mcp.shinzero.workers.dev/mcpIt automatically uses your DeepCrawl API key configured in the server. No additional setup required on your end.
Troubleshooting
If you encounter issues:
- Connection errors - Ensure your MCP client supports Streamable HTTP transport
- Rate limiting - The server has no rate limits, but be respectful of target websites
- Large pages - Use
get_markdownfor better performance on large pages
SDK vs MCP
| Feature | SDK | MCP |
|---|---|---|
| Setup required | Yes (API key) | No |
| Best for | Production apps | Quick prototyping, AI agents |
| Performance | Faster (direct) | Slightly slower (HTTP) |
| Caching | Configurable | Enabled by default |