# Brightdata

```json
{
  "name": "Brightdata",
  "slug": "brightdata",
  "url": "https://composio.dev/toolkits/brightdata",
  "markdown_url": "https://composio.dev/toolkits/brightdata.md",
  "logo_url": "https://logos.composio.dev/api/brightdata",
  "categories": [
    "data & analytics"
  ],
  "is_composio_managed": false,
  "updated_at": "2026-05-12T10:04:08.523Z"
}
```

![Brightdata logo](https://logos.composio.dev/api/brightdata)

## Description

Securely connect your AI agents and chatbots (Claude, ChatGPT, Cursor, etc) with Brightdata MCP or direct API to scrape websites, extract SERP results, bypass anti-bot protection, and automate web data collection through natural language.

## Summary

Brightdata is a leading web data platform offering advanced scraping, SERP APIs, and anti-bot tools. It lets you collect public web data at scale, bypassing blocks and friction.

## Categories

- data & analytics

## Toolkit Details

- Tools: 10

## Images

- Logo: https://logos.composio.dev/api/brightdata

## Authentication

- **Api Key**
  - Type: `api_key`
  - Description: Api Key authentication for Bright Data.
  - Setup:
    - Configure Api Key credentials for Bright Data.
    - Use the credentials when creating an auth config in Composio.

## Suggested Prompts

- Download all product listings from this ecommerce site
- Check crawl status for my recent job
- Perform Google SERP search for latest news
- List available web unlocker proxy zones

## Supported Tools

| Tool slug | Name | Description |
|---|---|---|
| `BRIGHTDATA_CRAWL_API` | Trigger Site Crawl | Tool to trigger an asynchronous site crawl job to extract content across multiple pages or entire domains. Returns a snapshot_id required by BRIGHTDATA_GET_SNAPSHOT_STATUS (poll until complete) and BRIGHTDATA_GET_SNAPSHOT_RESULTS (call only after completion; querying early yields empty or partial data). Use when you need to start a crawl for a given dataset and list of URLs. Large crawls can produce very large payloads — fetch results incrementally. |
| `BRIGHTDATA_DATASET_LIST` | Browse Available Scrapers | Tool to list all available pre-made scrapers (datasets) from Bright Data's marketplace. Use when you need to browse available data sources for structured scraping. |
| `BRIGHTDATA_FILTER_DATASET` | Filter Dataset | Tool to apply custom filter criteria to a marketplace dataset (BETA). Use after selecting a dataset to generate a filtered snapshot. |
| `BRIGHTDATA_GET_LIST_OF_AVAILABLE_CITIES` | Get Available Cities | Tool to get available static network cities for a given country. Use when you need to configure static proxy endpoints after selecting a country. |
| `BRIGHTDATA_GET_LIST_OF_AVAILABLE_COUNTRIES` | Get Available Countries | Tool to list available countries and their ISO 3166-1 alpha-2 codes. Use when you need to configure zones with valid country codes before provisioning proxies. |
| `BRIGHTDATA_GET_SNAPSHOT_RESULTS` | Download Scraped Data | Tool to retrieve the scraped data from a completed crawl job by snapshot ID. Only call after confirming the job is complete via BRIGHTDATA_GET_SNAPSHOT_STATUS — querying before completion yields empty or partial data. Use after triggering a crawl or filtering a dataset to download the collected data. |
| `BRIGHTDATA_GET_SNAPSHOT_STATUS` | Check Crawl Status | Tool to check the processing status of a crawl job using snapshot ID. Call before attempting to download results to ensure data collection is complete. |
| `BRIGHTDATA_LIST_WEB_UNLOCKER_ZONES` | List Unlocker Zones | Tool to list your configured Web Unlocker zones and proxy endpoints. Use to view available zones for web scraping and bot protection bypass. |
| `BRIGHTDATA_SERP_SEARCH` | SERP Search | Tool to perform SERP (Search Engine Results Page) searches across different search engines using Bright Data's SERP Scrape API. Use when you need to retrieve search results, trending topics, or competitive analysis data. This action submits an asynchronous request and returns a response ID for tracking. |
| `BRIGHTDATA_WEB_UNLOCKER` | Web Unlocker | Tool to bypass bot detection, captcha, and other anti-scraping measures to extract content from websites. Use when you need to scrape websites that block automated access or require JavaScript rendering. Some responses may still contain CAPTCHA challenge pages or incomplete HTML; inspect the response before retrying, and avoid aggressive retry loops. |

## Supported Triggers

None listed.

## Installation and MCP Setup

### Path 1: SDK Installation

#### Path 1, Step 1: Install Composio

Install the Composio SDK
```python
pip install composio_openai
```

```typescript
npm install @composio/openai
```

#### Path 1, Step 2: Initialize Composio and Create Tool Router Session

Import and initialize Composio client, then create a Tool Router session
```python
from openai import OpenAI
from composio import Composio
from composio_openai import OpenAIResponsesProvider

composio = Composio(provider=OpenAIResponsesProvider())
openai = OpenAI()
session = composio.create(user_id='your-user-id')
```

```typescript
import OpenAI from 'openai';
import { Composio } from '@composio/core';
import { OpenAIResponsesProvider } from '@composio/openai';

const composio = new Composio({
  provider: new OpenAIResponsesProvider(),
});
const openai = new OpenAI({});
const session = await composio.create('your-user-id');
```

#### Path 1, Step 3: Execute Brightdata Tools via Tool Router with Your Agent

Get tools from Tool Router session and execute Brightdata actions with your Agent
```python
tools = session.tools
response = openai.responses.create(
  model='gpt-4.1',
  tools=tools,
  input=[{
    'role': 'user',
    'content': 'Trigger a crawl to extract all product listings from example.com'
  }]
)
result = composio.provider.handle_tool_calls(
  response=response,
  user_id='your-user-id'
)
print(result)
```

```typescript
const tools = session.tools;
const response = await openai.responses.create({
  model: 'gpt-4.1',
  tools: tools,
  input: [{
    role: 'user',
    content: 'Trigger a crawl to extract all product listings from example.com'
  }],
});
const result = await composio.provider.handleToolCalls(
  'your-user-id',
  response.output
);
console.log(result);
```

### Path 2: MCP Server Setup

#### Path 2, Step 1: Install Composio

Install the Composio SDK and Claude Agent SDK
```python
pip install composio claude-agent-sdk
```

```typescript
npm install @composio/core ai @ai-sdk/openai @ai-sdk/mcp
```

#### Path 2, Step 2: Create Tool Router Session

Initialize the Composio client and create a Tool Router session
```python
from composio import Composio
from claude_agent_sdk import ClaudeSDKClient, ClaudeAgentOptions

composio = Composio(api_key='your-composio-api-key')
session = composio.create(user_id='your-user-id')
url = session.mcp.url
```

```typescript
import { Composio } from '@composio/core';

const composio = new Composio({ apiKey: 'your-api-key' });

console.log("Creating Tool Router session...");
const { mcp } = await composio.create('your-user-id');
console.log(`Tool Router session created: ${mcp.url}`);
```

#### Path 2, Step 3: Connect to AI Agent

Use the MCP server with your AI agent
```python
import asyncio

options = ClaudeAgentOptions(
    permission_mode='bypassPermissions',
    mcp_servers={
        'tool_router': {
            'type': 'http',
            'url': url,
            'headers': {
                'x-api-key': 'your-composio-api-key'
            }
        }
    },
    system_prompt='You are a helpful assistant with access to Brightdata tools.',
    max_turns=10
)

async def main():
    async with ClaudeSDKClient(options=options) as client:
        await client.query('Trigger a site crawl for https://example.com and download the scraped data')
        async for message in client.receive_response():
            if hasattr(message, 'content'):
                for block in message.content:
                    if hasattr(block, 'text'):
                        print(block.text)

asyncio.run(main())
```

```typescript
import { openai } from '@ai-sdk/openai';
import { experimental_createMCPClient as createMCPClient } from '@ai-sdk/mcp';
import { generateText, stepCountIs } from 'ai';

const client = await createMCPClient({
  transport: {
    type: 'http',
    url: mcp.url,
    headers: { 'x-api-key': 'your-composio-api-key' }
  }
});

const tools = await client.tools();

const { text } = await generateText({
  model: openai('gpt-4o'),
  tools,
  messages: [{ role: 'user', content: 'Trigger a site crawl for https://example.com and download the scraped data' }],
  stopWhen: stepCountIs(5)
});

console.log(`Agent: ${text}`);
```

## Why Use Composio?

### 1. AI Native Brightdata Integration

- Supports both Brightdata MCP and direct API based integrations
- Structured, LLM-friendly schemas for reliable tool execution
- Rich coverage for reading, writing, and querying your Brightdata data

### 2. Managed Auth

- Built-in API key management and automatic rotation
- Central place to manage, scope, and revoke Brightdata access
- Per user and per environment credentials instead of hard-coded keys

### 3. Agent Optimized Design

- Tools are tuned using real error and success rates to improve reliability over time
- Comprehensive execution logs so you always know what ran, when, and on whose behalf

### 4. Enterprise Grade Security

- Fine-grained RBAC so you control which agents and users can access Brightdata
- Scoped, least privilege access to Brightdata resources
- Full audit trail of agent actions to support review and compliance

## Use Brightdata with any AI Agent Framework

Choose a framework you want to connect Brightdata with:

- [ChatGPT](https://composio.dev/toolkits/brightdata/framework/chatgpt)
- [OpenAI Agents SDK](https://composio.dev/toolkits/brightdata/framework/open-ai-agents-sdk)
- [Claude Agent SDK](https://composio.dev/toolkits/brightdata/framework/claude-agents-sdk)
- [Claude Code](https://composio.dev/toolkits/brightdata/framework/claude-code)
- [Claude Cowork](https://composio.dev/toolkits/brightdata/framework/claude-cowork)
- [Codex](https://composio.dev/toolkits/brightdata/framework/codex)
- [Cursor](https://composio.dev/toolkits/brightdata/framework/cursor)
- [VS Code](https://composio.dev/toolkits/brightdata/framework/vscode)
- [OpenCode](https://composio.dev/toolkits/brightdata/framework/opencode)
- [OpenClaw](https://composio.dev/toolkits/brightdata/framework/openclaw)
- [Hermes](https://composio.dev/toolkits/brightdata/framework/hermes-agent)
- [Google ADK](https://composio.dev/toolkits/brightdata/framework/google-adk)
- [LangChain](https://composio.dev/toolkits/brightdata/framework/langchain)
- [Vercel AI SDK](https://composio.dev/toolkits/brightdata/framework/ai-sdk)
- [Mastra AI](https://composio.dev/toolkits/brightdata/framework/mastra-ai)
- [LlamaIndex](https://composio.dev/toolkits/brightdata/framework/llama-index)
- [CrewAI](https://composio.dev/toolkits/brightdata/framework/crew-ai)
- [Pydantic AI](https://composio.dev/toolkits/brightdata/framework/pydantic-ai)
- [AutoGen](https://composio.dev/toolkits/brightdata/framework/autogen)

## Related Toolkits

- [Excel](https://composio.dev/toolkits/excel) - Microsoft Excel is a robust spreadsheet application for organizing, analyzing, and visualizing data. It's the go-to tool for calculations, reporting, and flexible data management.
- [21risk](https://composio.dev/toolkits/_21risk) - 21RISK is a web app built for easy checklist, audit, and compliance management. It streamlines risk processes so teams can focus on what matters.
- [Abstract](https://composio.dev/toolkits/abstract) - Abstract provides a suite of APIs for automating data validation and enrichment tasks. It helps developers streamline workflows and ensure data quality with minimal effort.
- [Addressfinder](https://composio.dev/toolkits/addressfinder) - Addressfinder is a data quality platform for verifying addresses, emails, and phone numbers. It helps you ensure accurate customer and contact data every time.
- [Agentql](https://composio.dev/toolkits/agentql) - Agentql is a toolkit that connects AI agents to the web using a specialized query language. It enables structured web interaction and data extraction for smarter automations.
- [Agenty](https://composio.dev/toolkits/agenty) - Agenty is a web scraping and automation platform for extracting data and automating browser tasks—no coding needed. It streamlines data collection, monitoring, and repetitive online actions.
- [Ambee](https://composio.dev/toolkits/ambee) - Ambee is an environmental data platform providing real-time, hyperlocal APIs for air quality, weather, and pollen. Get precise environmental insights to power smarter decisions in your apps and workflows.
- [Ambient weather](https://composio.dev/toolkits/ambient_weather) - Ambient Weather is a platform for personal weather stations with a robust API for accessing local, real-time, and historical weather data. Get detailed environmental insights directly from your own sensors for smarter apps and automations.
- [Anonyflow](https://composio.dev/toolkits/anonyflow) - Anonyflow is a service for encryption-based data anonymization and secure data sharing. It helps organizations meet GDPR, CCPA, and HIPAA data privacy compliance requirements.
- [Api ninjas](https://composio.dev/toolkits/api_ninjas) - Api ninjas offers 120+ public APIs spanning categories like weather, finance, sports, and more. Developers use it to supercharge apps with real-time data and actionable endpoints.
- [Api sports](https://composio.dev/toolkits/api_sports) - Api sports is a comprehensive sports data platform covering 2,000+ competitions with live scores and 15+ years of stats. Instantly access up-to-date sports information for analysis, apps, or chatbots.
- [Apify](https://composio.dev/toolkits/apify) - Apify is a cloud platform for building, deploying, and managing web scraping and automation tools called Actors. It lets you automate data extraction and workflow tasks at scale—no infrastructure headaches.
- [Autom](https://composio.dev/toolkits/autom) - Autom is a lightning-fast search engine results data platform for Google, Bing, and Brave. Developers use it to access fresh, low-latency SERP data on demand.
- [Beaconchain](https://composio.dev/toolkits/beaconchain) - Beaconchain is a real-time analytics platform for Ethereum 2.0's Beacon Chain. It provides detailed insights into validators, blocks, and overall network performance.
- [Big data cloud](https://composio.dev/toolkits/big_data_cloud) - BigDataCloud provides APIs for geolocation, reverse geocoding, and address validation. Instantly access reliable location intelligence to enhance your applications and workflows.
- [Bigpicture io](https://composio.dev/toolkits/bigpicture_io) - BigPicture.io offers APIs for accessing detailed company and profile data. Instantly enrich your applications with up-to-date insights on 20M+ businesses.
- [Bitquery](https://composio.dev/toolkits/bitquery) - Bitquery is a blockchain data platform offering indexed, real-time, and historical data from 40+ blockchains via GraphQL APIs. Get unified, reliable access to complex on-chain data for analytics, trading, and research.
- [Builtwith](https://composio.dev/toolkits/builtwith) - BuiltWith is a web technology profiler that uncovers the technologies powering any website. Gain actionable insights into analytics, hosting, and content management stacks for smarter research and lead generation.
- [Byteforms](https://composio.dev/toolkits/byteforms) - Byteforms is an all-in-one platform for creating forms, managing submissions, and integrating data. It streamlines workflows by centralizing form data collection and automation.
- [Cabinpanda](https://composio.dev/toolkits/cabinpanda) - Cabinpanda is a data collection platform for building and managing online forms. It helps streamline how you gather, organize, and analyze responses.

## Frequently Asked Questions

### Do I need my own developer credentials to use Brightdata with Composio?

Yes, Brightdata requires you to configure your own API key credentials. Once set up, Composio handles secure credential storage and API request handling for you.

### Can I use multiple toolkits together?

Yes! Composio's Tool Router enables agents to use multiple toolkits. [Learn more](https://docs.composio.dev/tool-router/overview).

### Is Composio secure?

Composio is SOC 2 and ISO 27001 compliant with all data encrypted in transit and at rest. [Learn more](https://trust.composio.dev).

### What if the API changes?

Composio maintains and updates all toolkit integrations automatically, so your agents always work with the latest API versions.

---
[See all toolkits](https://composio.dev/toolkits) · [Composio docs](https://docs.composio.dev/llms.txt)
