# How to integrate Scrape do MCP with Hermes

```json
{
  "title": "How to integrate Scrape do MCP with Hermes",
  "toolkit": "Scrape do",
  "toolkit_slug": "scrape_do",
  "framework": "Hermes",
  "framework_slug": "hermes-agent",
  "url": "https://composio.dev/toolkits/scrape_do/framework/hermes-agent",
  "markdown_url": "https://composio.dev/toolkits/scrape_do/framework/hermes-agent.md",
  "updated_at": "2026-05-06T08:26:50.052Z"
}
```

## Introduction

Hermes is a 24/7 autonomous agent that lives on your computer or server — it remembers what it learns and evolves as your usage grows.
This guide explains the easiest and most robust way to connect your Scrape do account to Hermes. You can do this through either Composio Connect CLI or Composio Connect MCP. For personal use we recommend the CLI, but you won't go wrong with MCP either.

## Also integrate Scrape do with

- [OpenAI Agents SDK](https://composio.dev/toolkits/scrape_do/framework/open-ai-agents-sdk)
- [Claude Agent SDK](https://composio.dev/toolkits/scrape_do/framework/claude-agents-sdk)
- [Claude Code](https://composio.dev/toolkits/scrape_do/framework/claude-code)
- [Claude Cowork](https://composio.dev/toolkits/scrape_do/framework/claude-cowork)
- [Codex](https://composio.dev/toolkits/scrape_do/framework/codex)
- [OpenClaw](https://composio.dev/toolkits/scrape_do/framework/openclaw)
- [CLI](https://composio.dev/toolkits/scrape_do/framework/cli)
- [Google ADK](https://composio.dev/toolkits/scrape_do/framework/google-adk)
- [LangChain](https://composio.dev/toolkits/scrape_do/framework/langchain)
- [Vercel AI SDK](https://composio.dev/toolkits/scrape_do/framework/ai-sdk)
- [Mastra AI](https://composio.dev/toolkits/scrape_do/framework/mastra-ai)
- [LlamaIndex](https://composio.dev/toolkits/scrape_do/framework/llama-index)
- [CrewAI](https://composio.dev/toolkits/scrape_do/framework/crew-ai)

## TL;DR

### What is Composio Connect?
Composio Connect is a consumer offering that lets anyone plug 1,000+ applications directly into their agent harness — including Hermes. It can:
- Search and load tools from relevant toolkits on-demand, reducing context usage.
- Chain multiple tools to accomplish complex workflows via a remote workbench, without excessive back-and-forth with the LLM.
- Manage app authentication end-to-end with zero manual overhead.

## Connect Scrape do to Hermes

### Integrating Scrape do with Hermes
### Using Composio Connect CLI
1. Install the Composio CLI
Run the install script directly, or paste https://composio.dev/hermes into your Hermes chat box to have it installed for you.

```bash
curl -fsSL https://composio.dev/install | bash
```

## What is the Scrape do MCP server, and what's possible with it?

The Scrape do MCP server is an implementation of the Model Context Protocol that connects your AI agent and assistants like Claude, Cursor, etc directly to your Scrape do account. It provides structured and secure access to robust web scraping tools, so your agent can perform actions like scraping dynamic pages, managing sessions, setting custom headers or proxies, and extracting structured data from any website on your behalf.
- Dynamic page scraping with headless browsers: Retrieve fully rendered HTML content from JavaScript-heavy or protected websites by leveraging advanced browser emulation and proxy rotation.
- Custom scraping session management: Set device type, cookies, wait times, and custom headers to imitate different users, maintain sessions, or access device-specific content for tailored data extraction.
- Proxy and anti-bot bypass control: Enable super or proxy modes to utilize residential, mobile, or datacenter proxies, helping your agent bypass strict anti-bot systems and geo-restrictions seamlessly.
- Targeted resource filtering: Block specific URLs like ads or analytics scripts during scraping to increase speed, avoid distractions, and improve privacy.
- Account usage and statistics retrieval: Access real-time usage stats, subscription status, and remaining request limits so your agent can monitor scraping quotas and avoid interruptions.

## Supported Tools

| Tool slug | Name | Description |
|---|---|---|
| `SCRAPE_DO_GET_ACCOUNT_INFO` | Get Account Information | Retrieves account information and usage statistics from scrape.do. this action makes a get request to the scrape.do info endpoint to fetch: - subscription status - concurrent request limits and usage - monthly request limits and remaining requests - real-time usage statistics rate limit: maximum 10 requests per minute |
| `SCRAPE_DO_GET_RENDER_PAGE` | Get rendered page content | This tool allows you to scrape web pages with javascript rendering enabled. it's particularly useful for scraping dynamic websites where content is loaded through javascript. the tool will wait for the javascript to execute and return the fully rendered html content. |
| `SCRAPE_DO_SCRAPE_DO_GET_PAGE` | Scrape webpage using scrape.do | A tool to scrape web pages using scrape.do's api service. it makes a basic get request to fetch the content of a target webpage while handling anti-bot protections and proxy rotation automatically. |
| `SCRAPE_DO_SCRAPE_DO_PROXY_MODE` | Use Scrape.do Proxy Mode | This tool implements the proxy mode functionality of scrape.do, which allows routing requests through their proxy server. it provides an alternative way to access web scraping capabilities by handling complex javascript-rendered pages, geolocation-based routing, device simulation, and built-in anti-bot and retry mechanisms. |
| `SCRAPE_DO_SCRAPE_DO_SET_COOKIES` | Set Cookies for Scraping | This tool allows users to set specific cookies for their scraping requests to a target website. it is useful for maintaining session states or authentication through cookies. |
| `SCRAPE_DO_SCRAPE_DO_SET_SUPER_MODE` | Set Scrape.do Super Mode | The scrape do set super mode tool enables enhanced scraping by using residential and mobile proxies, bypassing blocks and restrictions associated with datacenter ips. when the 'super' parameter is set to true, it activates a mode that leverages a network of residential ip addresses, which is particularly useful to bypass strict anti-bot measures and for accessing websites that block datacenter ips. |
| `SCRAPE_DO_SET_BLOCK_URLS` | Block specific URLs during scraping | This tool allows users to block specific urls during the scraping process. it's particularly useful for blocking unwanted resources like analytics scripts, advertisements, or any other urls that might interfere with the scraping process or slow it down. it provides granular control by allowing users to specify url patterns to block, thereby improving scraping performance and maintaining privacy. |
| `SCRAPE_DO_SET_CUSTOM_HEADERS` | Set custom headers for scrape.do request | A tool to send custom headers with scrape.do requests. this allows simulating specific browser behaviors or adding authentication headers by controlling all headers sent to the target website. |
| `SCRAPE_DO_SET_CUSTOM_WAIT` | Set Custom Wait Time | This tool sets the custom wait time in milliseconds after page load when using the render option in scrape.do. it is particularly useful for dealing with dynamic content to ensure that it is fully loaded before scraping, especially on javascript-heavy websites or single-page applications. the action allows fine-tuned control over the rendering wait time and must be used with render=true. |
| `SCRAPE_DO_SET_DEVICE_TYPE` | Set Device Type for Scraping | This tool allows users to set the device type (desktop, mobile, or tablet) for making scraping requests. it is used to emulate different devices, which helps in testing responsive designs or fetching device-specific content. |
| `SCRAPE_DO_SET_DISABLE_REDIRECTION` | Set Disable Redirection | Controls the automatic redirection behavior of scrape.do requests. when enabled (disable redirection=true), prevents the automatic following of redirects during web scraping operations. this allows: - inspection of the redirect chain - capturing intermediate redirect responses - manual control of redirection flow - analysis of http status codes of redirect responses the redirect url will be available in the scrape.do-target-redirected-location response header. |
| `SCRAPE_DO_SET_PURE_COOKIES` | Set Pure Cookies Mode | This tool enables getting the original set-cookie headers from target websites instead of the processed scrape.do-cookies format. when enabled, this parameter returns the original set-cookie headers from the target website rather than using the default scrape.do-cookies header format. |
| `SCRAPE_DO_SET_REGIONAL_GEO_CODE` | Set Regional Geolocation for Scraping | This tool allows users to set a broader geographical targeting by specifying a region code instead of a specific country code. this is useful when you want to scrape content from an entire region rather than a specific country. note that this feature requires super mode to be enabled and is only available for business plan or higher subscriptions. |
| `SCRAPE_DO_SET_RETRY_TIMEOUT` | Set Retry Timeout | This tool allows users to set the maximum wait time (in milliseconds) before retrying a failed request in scrape.do. it requires a parameter 'retry timeout' (integer) which specifies the maximum time to wait before retrying, with a default of 15000 ms. it is designed to improve the reliability of web scraping operations, especially when dealing with unstable or slow-responding websites. |
| `SCRAPE_DO_SET_SCREENSHOT` | Set Screenshot Capture for Scraping | This tool enables the screenshot functionality for the scrape.do api, allowing users to capture a visual representation of the scraped webpage. when enabled, the api will return a screenshot of the rendered page along with the regular response. features: - basic screenshot capture - full page screenshot capture - capture specific area using css selector |
| `SCRAPE_DO_SET_SESSION_ID` | Set Session ID for Sticky Sessions | This tool implements the session id functionality for scrape.do to maintain a sticky session with the same proxy ip across multiple requests. it achieves this by adding a sessionid parameter to the query parameters of any scraping request, which is crucial for ensuring session consistency when scraping websites with stringent session requirements. |
| `SCRAPE_DO_SET_WAIT_FOR_SELECTOR` | Set Wait For Selector | This action allows setting a css selector to wait for before considering the page load complete. it is particularly useful when scraping javascript-heavy pages to ensure that certain elements have loaded dynamically. |
| `SCRAPE_DO_SET_WAIT_UNTIL` | Set Wait Until Condition | This tool sets the waituntil parameter for the scrape.do api, defining when the rendering should consider the page loaded during javascript execution. it is particularly useful for handling dynamic websites by specifying conditions such as 'domcontentloaded', 'networkidle0', or 'networkidle2'. |
| `SCRAPE_DO_SHOW_WEBSOCKET_REQUESTS` | Monitor WebSocket requests using scrape.do | This tool provides the ability to view websocket requests made by a webpage. it requires using render=true and returnjson=true parameters along with showwebsocketrequests=true to enable logging of websocket requests. |

## Supported Triggers

None listed.

## Creating MCP Server - Stand-alone vs Composio SDK

The Scrape do MCP server provides comprehensive access to Scrape do operations through Composio. Once connected, Hermes can perform all major Scrape do actions on your behalf using natural language commands.

## Complete Code

None listed.

## Conclusion

### Way Forward
With Scrape do connected, Hermes can now act on your behalf whenever it detects a relevant task or you ask it to.
From here, you can extend Hermes further:
- Connect more apps: Calendar, Slack, Notion, Linear, and hundreds of others are available through the same Composio Connect setup. Each new integration compounds what Hermes can do for you.
- Build workflows across tools: Once multiple apps are connected, Hermes can chain actions together — turn an email into a calendar invite, a Slack message into a Linear ticket, or a meeting note into a follow-up draft.
- Let it learn your patterns: The more you use Hermes, the better it gets at anticipating how you'd handle recurring tasks. Give it feedback on drafts and decisions, and it will adapt.
If you run into trouble or want to share what you've built, join the [community](https://discord.com/invite/composio) or check out the [Docs](https://docs.composio.dev?utm_source=toolkits&utm_medium=framework_template&utm_campaign=hermes&utm_content=docs) for deeper configuration options.

## How to build Scrape do MCP Agent with another framework

- [OpenAI Agents SDK](https://composio.dev/toolkits/scrape_do/framework/open-ai-agents-sdk)
- [Claude Agent SDK](https://composio.dev/toolkits/scrape_do/framework/claude-agents-sdk)
- [Claude Code](https://composio.dev/toolkits/scrape_do/framework/claude-code)
- [Claude Cowork](https://composio.dev/toolkits/scrape_do/framework/claude-cowork)
- [Codex](https://composio.dev/toolkits/scrape_do/framework/codex)
- [OpenClaw](https://composio.dev/toolkits/scrape_do/framework/openclaw)
- [CLI](https://composio.dev/toolkits/scrape_do/framework/cli)
- [Google ADK](https://composio.dev/toolkits/scrape_do/framework/google-adk)
- [LangChain](https://composio.dev/toolkits/scrape_do/framework/langchain)
- [Vercel AI SDK](https://composio.dev/toolkits/scrape_do/framework/ai-sdk)
- [Mastra AI](https://composio.dev/toolkits/scrape_do/framework/mastra-ai)
- [LlamaIndex](https://composio.dev/toolkits/scrape_do/framework/llama-index)
- [CrewAI](https://composio.dev/toolkits/scrape_do/framework/crew-ai)

## Related Toolkits

- [Excel](https://composio.dev/toolkits/excel) - Microsoft Excel is a robust spreadsheet application for organizing, analyzing, and visualizing data. It's the go-to tool for calculations, reporting, and flexible data management.
- [21risk](https://composio.dev/toolkits/_21risk) - 21RISK is a web app built for easy checklist, audit, and compliance management. It streamlines risk processes so teams can focus on what matters.
- [Abstract](https://composio.dev/toolkits/abstract) - Abstract provides a suite of APIs for automating data validation and enrichment tasks. It helps developers streamline workflows and ensure data quality with minimal effort.
- [Addressfinder](https://composio.dev/toolkits/addressfinder) - Addressfinder is a data quality platform for verifying addresses, emails, and phone numbers. It helps you ensure accurate customer and contact data every time.
- [Agentql](https://composio.dev/toolkits/agentql) - Agentql is a toolkit that connects AI agents to the web using a specialized query language. It enables structured web interaction and data extraction for smarter automations.
- [Agenty](https://composio.dev/toolkits/agenty) - Agenty is a web scraping and automation platform for extracting data and automating browser tasks—no coding needed. It streamlines data collection, monitoring, and repetitive online actions.
- [Ambee](https://composio.dev/toolkits/ambee) - Ambee is an environmental data platform providing real-time, hyperlocal APIs for air quality, weather, and pollen. Get precise environmental insights to power smarter decisions in your apps and workflows.
- [Ambient weather](https://composio.dev/toolkits/ambient_weather) - Ambient Weather is a platform for personal weather stations with a robust API for accessing local, real-time, and historical weather data. Get detailed environmental insights directly from your own sensors for smarter apps and automations.
- [Anonyflow](https://composio.dev/toolkits/anonyflow) - Anonyflow is a service for encryption-based data anonymization and secure data sharing. It helps organizations meet GDPR, CCPA, and HIPAA data privacy compliance requirements.
- [Api ninjas](https://composio.dev/toolkits/api_ninjas) - Api ninjas offers 120+ public APIs spanning categories like weather, finance, sports, and more. Developers use it to supercharge apps with real-time data and actionable endpoints.
- [Api sports](https://composio.dev/toolkits/api_sports) - Api sports is a comprehensive sports data platform covering 2,000+ competitions with live scores and 15+ years of stats. Instantly access up-to-date sports information for analysis, apps, or chatbots.
- [Apify](https://composio.dev/toolkits/apify) - Apify is a cloud platform for building, deploying, and managing web scraping and automation tools called Actors. It lets you automate data extraction and workflow tasks at scale—no infrastructure headaches.
- [Autom](https://composio.dev/toolkits/autom) - Autom is a lightning-fast search engine results data platform for Google, Bing, and Brave. Developers use it to access fresh, low-latency SERP data on demand.
- [Beaconchain](https://composio.dev/toolkits/beaconchain) - Beaconchain is a real-time analytics platform for Ethereum 2.0's Beacon Chain. It provides detailed insights into validators, blocks, and overall network performance.
- [Big data cloud](https://composio.dev/toolkits/big_data_cloud) - BigDataCloud provides APIs for geolocation, reverse geocoding, and address validation. Instantly access reliable location intelligence to enhance your applications and workflows.
- [Bigpicture io](https://composio.dev/toolkits/bigpicture_io) - BigPicture.io offers APIs for accessing detailed company and profile data. Instantly enrich your applications with up-to-date insights on 20M+ businesses.
- [Bitquery](https://composio.dev/toolkits/bitquery) - Bitquery is a blockchain data platform offering indexed, real-time, and historical data from 40+ blockchains via GraphQL APIs. Get unified, reliable access to complex on-chain data for analytics, trading, and research.
- [Brightdata](https://composio.dev/toolkits/brightdata) - Brightdata is a leading web data platform offering advanced scraping, SERP APIs, and anti-bot tools. It lets you collect public web data at scale, bypassing blocks and friction.
- [Builtwith](https://composio.dev/toolkits/builtwith) - BuiltWith is a web technology profiler that uncovers the technologies powering any website. Gain actionable insights into analytics, hosting, and content management stacks for smarter research and lead generation.
- [Byteforms](https://composio.dev/toolkits/byteforms) - Byteforms is an all-in-one platform for creating forms, managing submissions, and integrating data. It streamlines workflows by centralizing form data collection and automation.

## Frequently Asked Questions

### What are the differences in Tool Router MCP and Scrape do MCP?

With a standalone Scrape do MCP server, the agents and LLMs can only access a fixed set of Scrape do tools tied to that server. However, with the Composio Tool Router, agents can dynamically load tools from Scrape do and many other apps based on the task at hand, all through a single MCP endpoint.

### Can I use Tool Router MCP with Hermes?

Yes, you can. Hermes fully supports MCP integration. You get structured tool calling, message history handling, and model orchestration while Tool Router takes care of discovering and serving the right Scrape do tools.

### Can I manage the permissions and scopes for Scrape do while using Tool Router?

Yes, absolutely. You can configure which Scrape do scopes and actions are allowed when connecting your account to Composio. You can also bring your own OAuth credentials or API configuration so you keep full control over what the agent can do.

### How safe is my data with Composio Tool Router?

All sensitive data such as tokens, keys, and configuration is fully encrypted at rest and in transit. Composio is SOC 2 Type 2 compliant and follows strict security practices so your Scrape do data and credentials are handled as safely as possible.

---
[See all toolkits](https://composio.dev/toolkits) · [Composio docs](https://docs.composio.dev/llms.txt)
