# Firecrawl

```json
{
  "name": "Firecrawl",
  "slug": "firecrawl",
  "url": "https://composio.dev/toolkits/firecrawl",
  "markdown_url": "https://composio.dev/toolkits/firecrawl.md",
  "logo_url": "https://logos.composio.dev/api/firecrawl",
  "categories": [
    "analytics & data"
  ],
  "is_composio_managed": false,
  "updated_at": "2026-05-12T10:11:40.863Z"
}
```

![Firecrawl logo](https://logos.composio.dev/api/firecrawl)

## Description

Securely connect your AI agents and chatbots (Claude, ChatGPT, Cursor, etc) with Firecrawl MCP or direct API to crawl websites, extract structured data, index web content, and analyze online sources through natural language.

## Summary

Firecrawl automates large-scale web crawling and data extraction. It helps organizations efficiently gather, index, and analyze content from online sources.

## Categories

- analytics & data

## Toolkit Details

- Tools: 29

## Images

- Logo: https://logos.composio.dev/api/firecrawl

## Authentication

- **Api Key**
  - Type: `api_key`
  - Description: Api Key authentication for Firecrawl.
  - Setup:
    - Configure Api Key credentials for Firecrawl.
    - Use the credentials when creating an auth config in Composio.

## Suggested Prompts

- Extract all product prices from this e-commerce site
- Crawl competitor blogs for latest article summaries
- Map all subpages linked from homepage URL
- Search for recent news articles about AI trends

## Supported Tools

| Tool slug | Name | Description |
|---|---|---|
| `FIRECRAWL_AGENT_CANCEL` | Cancel an agent job | Tool to cancel an in-progress agent job by its ID. Use when you need to terminate an active agent operation. The API returns a success boolean upon cancellation. |
| `FIRECRAWL_BATCH_SCRAPE` | Batch scrape multiple URLs | Tool to scrape multiple URLs in batch with concurrent processing. Use when you need to scrape multiple web pages efficiently with customizable formats and content filtering. |
| `FIRECRAWL_BATCH_SCRAPE_CANCEL` | Cancel a batch scrape job | Tool to cancel a running batch scrape job using its unique identifier. Use when you need to terminate an in-progress batch scrape operation. |
| `FIRECRAWL_BATCH_SCRAPE_GET` | Get batch scrape status | Retrieves the current status and results of a batch scrape job using the job ID. Use this to check batch scrape progress and retrieve scraped data. |
| `FIRECRAWL_BATCH_SCRAPE_GET_ERRORS` | Get errors from batch scrape job | Tool to retrieve error details from a batch scrape job, including failed URLs and URLs blocked by robots.txt. Use when you need to debug or understand why certain pages failed to scrape in a batch operation. |
| `FIRECRAWL_CRAWL` | Start a web crawl | Initiates a Firecrawl web crawl from a given URL, applying various filtering and content extraction rules, and polls until the job is complete; ensure the URL is accessible and any regex patterns for paths are valid. |
| `FIRECRAWL_CANCEL_A_CRAWL_JOB` | Cancel a crawl job | Cancels an active or queued web crawl job using its ID; attempting to cancel completed, failed, or previously canceled jobs will not change their state. |
| `FIRECRAWL_CANCEL_A_CRAWL_JOB` | Cancel a crawl job | Tool to cancel a running crawl job by its ID. Use when you need to stop an active crawl operation. The API returns a status of 'cancelled' upon successful cancellation. |
| `FIRECRAWL_CRAWL_GET` | Get crawl job status | Tool to retrieve the status and results of a Firecrawl crawl job. Use when you need to check the progress or get data from an ongoing or completed crawl operation. Returns crawl status, progress metrics, credits used, and the crawled page data. |
| `FIRECRAWL_CRAWL_GET_ERRORS` | Get errors from a crawl job | Tool to retrieve errors from a Firecrawl crawl job. Use when you need to understand why certain pages failed to scrape or which URLs were blocked by robots.txt during a crawl operation. |
| `FIRECRAWL_CRAWL_LIST_ACTIVE` | Get all active crawl jobs | Tool to retrieve all active crawl jobs for the authenticated team. Use when you need to see which crawl operations are currently running. |
| `FIRECRAWL_CRAWL_PARAMS_PREVIEW` | Preview crawl parameters | Preview crawl parameters before starting a crawl by generating optimal configuration from natural language instructions. Use this tool to understand what crawl settings will be applied based on your requirements before executing a full crawl operation. The endpoint intelligently interprets natural language prompts to configure crawl parameters like include/exclude paths, depth limits, and domain scope. |
| `FIRECRAWL_CRAWL_V2` | Start a web crawl (v2) [NEW] | [NEW v2 API] Initiates a Firecrawl v2 web crawl with enhanced features over v1: natural language prompts for automatic crawler configuration, crawlEntireDomain for sibling/parent page discovery, better depth control with maxDiscoveryDepth, subdomain support, and full webhook configuration. Polls until crawl is complete. |
| `FIRECRAWL_CREDIT_USAGE_GET` | Get team credit usage | Tool to get current team credit usage information. Use when you need to check remaining credits or billing period details. |
| `FIRECRAWL_CREDIT_USAGE_GET_HISTORICAL` | Get historical team credit usage | Tool to retrieve historical team credit usage on a monthly basis. Use when you need to analyze credit consumption patterns over time, optionally segmented by API key. |
| `FIRECRAWL_EXTRACT` | Extract structured data | Extracts structured data from web pages by initiating an extraction job and polling for completion; requires a natural language `prompt` or a JSON `schema` (one must be provided). |
| `FIRECRAWL_EXTRACT_GET` | Get extract job status | Tool to retrieve the status and results of a previously submitted extract job. Use when you need to check the progress or get the final results of an extraction operation. |
| `FIRECRAWL_GET_AGENT_STATUS` | Get agent job status | Tool to get the status and results of an agent job. Use when you need to check if an agent job has completed and retrieve the collected data. Agent jobs autonomously search, navigate, and extract data from the web. |
| `FIRECRAWL_GET_DEEP_RESEARCH_STATUS` | Get deep research status | Retrieves the status and results of a deep research job by its ID. Use when you need to check the progress or retrieve the final analysis of a deep research operation. |
| `FIRECRAWL_GET_THE_STATUS_OF_A_CRAWL_JOB` | Get the status of a crawl job | Retrieves the current status, progress, and details of a web crawl job, using the job ID obtained when the crawl was initiated. |
| `FIRECRAWL_LLMS_TXT_GENERATE` | Generate LLMs.txt for a website | Initiates an async job to generate an LLMs.txt file for a website, converting web content into LLM-friendly format. Returns a job ID to check status and retrieve results. Use when you need to create a standardized, machine-readable representation of website content for language models. |
| `FIRECRAWL_LLMS_TXT_GET` | Get LLMs.txt generation job status | Tool to get the status and results of an LLMs.txt generation job. Use when you need to check if a job has completed and retrieve the generated content. |
| `FIRECRAWL_MAP_MULTIPLE_URLS_BASED_ON_OPTIONS` | Map multiple URLs | Maps a website by discovering URLs from a starting base URL, with options to customize the crawl via search query, subdomain inclusion, sitemap handling, and result limits; search effectiveness is site-dependent. |
| `FIRECRAWL_QUEUE_GET` | Get team queue status | Tool to retrieve metrics about the team's scrape queue. Use when you need to check queue status, job counts, or concurrency limits. |
| `FIRECRAWL_SCRAPE` | Scrape URL | Scrapes a publicly accessible URL, optionally performing pre-scrape browser actions or extracting structured JSON using an LLM, to retrieve content in specified formats. |
| `FIRECRAWL_SEARCH` | Search | Performs a web search for a query, scrapes content from the top search results using Firecrawl, and returns details in specified formats. |
| `FIRECRAWL_START_AGENT` | Start an agent job | Tool to start an agent job for agentic web extraction with multi-page navigation and interaction capabilities. Use when you need to autonomously gather data from the web with complex navigation requirements. The agent can search, navigate, and extract information across multiple pages based on your natural language prompt. |
| `FIRECRAWL_TOKEN_USAGE_GET` | Get team token usage | Tool to retrieve the current team's token usage and balance information for Firecrawl's Extract feature. Use when you need to check remaining token credits, plan allocation, or billing period details. |
| `FIRECRAWL_TOKEN_USAGE_GET_HISTORICAL` | Get historical team token usage | Tool to retrieve historical team token usage on a monthly basis. Use when you need to analyze token consumption patterns over time, optionally segmented by API key. |

## Supported Triggers

None listed.

## Installation and MCP Setup

### Path 1: SDK Installation

#### Path 1, Step 1: Install Composio

Install the Composio SDK
```python
pip install composio_openai
```

```typescript
npm install @composio/openai
```

#### Path 1, Step 2: Initialize Composio and Create Tool Router Session

Import and initialize Composio client, then create a Tool Router session
```python
from openai import OpenAI
from composio import Composio
from composio_openai import OpenAIResponsesProvider

composio = Composio(provider=OpenAIResponsesProvider())
openai = OpenAI()
session = composio.create(user_id='your-user-id')
```

```typescript
import OpenAI from 'openai';
import { Composio } from '@composio/core';
import { OpenAIResponsesProvider } from '@composio/openai';

const composio = new Composio({
  provider: new OpenAIResponsesProvider(),
});
const openai = new OpenAI({});
const session = await composio.create('your-user-id');
```

#### Path 1, Step 3: Execute Firecrawl Tools via Tool Router with Your Agent

Get tools from Tool Router session and execute Firecrawl actions with your Agent
```python
tools = session.tools
response = openai.responses.create(
  model='gpt-4.1',
  tools=tools,
  input=[{
    'role': 'user',
    'content': 'Crawl https://docs.python.org and extract all tutorial links.'
  }]
)
result = composio.provider.handle_tool_calls(
  response=response,
  user_id='your-user-id'
)
print(result)
```

```typescript
const tools = session.tools;
const response = await openai.responses.create({
  model: 'gpt-4.1',
  tools: tools,
  input: [{
    role: 'user',
    content: 'Crawl https://docs.python.org and extract all tutorial links.'
  }],
});
const result = await composio.provider.handleToolCalls(
  'your-user-id',
  response.output
);
console.log(result);
```

### Path 2: MCP Server Setup

#### Path 2, Step 1: Install Composio

Install the Composio SDK and Claude Agent SDK
```python
pip install composio claude-agent-sdk
```

```typescript
npm install @composio/core ai @ai-sdk/openai @ai-sdk/mcp
```

#### Path 2, Step 2: Create Tool Router Session

Initialize the Composio client and create a Tool Router session
```python
from composio import Composio
from claude_agent_sdk import ClaudeSDKClient, ClaudeAgentOptions

composio = Composio(api_key='your-composio-api-key')
session = composio.create(user_id='your-user-id')
url = session.mcp.url
```

```typescript
import { Composio } from '@composio/core';

const composio = new Composio({ apiKey: 'your-api-key' });

console.log("Creating Tool Router session...");
const { mcp } = await composio.create('your-user-id');
console.log(`Tool Router session created: ${mcp.url}`);
```

#### Path 2, Step 3: Connect to AI Agent

Use the MCP server with your AI agent
```python
import asyncio

options = ClaudeAgentOptions(
    permission_mode='bypassPermissions',
    mcp_servers={
        'tool_router': {
            'type': 'http',
            'url': url,
            'headers': {
                'x-api-key': 'your-composio-api-key'
            }
        }
    },
    system_prompt='You are a helpful assistant with access to Firecrawl tools.',
    max_turns=10
)

async def main():
    async with ClaudeSDKClient(options=options) as client:
        await client.query('Extract structured data from https://docs.firecrawl.dev using schema extraction')
        async for message in client.receive_response():
            if hasattr(message, 'content'):
                for block in message.content:
                    if hasattr(block, 'text'):
                        print(block.text)

asyncio.run(main())
```

```typescript
import { openai } from '@ai-sdk/openai';
import { experimental_createMCPClient as createMCPClient } from '@ai-sdk/mcp';
import { generateText, stepCountIs } from 'ai';

const client = await createMCPClient({
  transport: {
    type: 'http',
    url: mcp.url,
    headers: { 'x-api-key': 'your-composio-api-key' }
  }
});

const tools = await client.tools();

const { text } = await generateText({
  model: openai('gpt-4o'),
  tools,
  messages: [{ role: 'user', content: 'Extract structured data from https://docs.firecrawl.dev using schema extraction' }],
  stopWhen: stepCountIs(5)
});

console.log(`Agent: ${text}`);
```

## Why Use Composio?

### 1. AI Native Firecrawl Integration

- Supports both Firecrawl MCP and direct API based integrations
- Structured, LLM-friendly schemas for reliable tool execution
- Rich coverage for crawling, extracting, and indexing web data

### 2. Managed Auth

- Built-in API key management with secure storage
- Central place to manage, scope, and revoke Firecrawl access
- Per user and per environment credentials instead of hard-coded keys

### 3. Agent Optimized Design

- Tools are tuned using real error and success rates to improve reliability over time
- Comprehensive execution logs so you always know what ran, when, and on whose behalf

### 4. Enterprise Grade Security

- Fine-grained RBAC so you control which agents and users can access Firecrawl
- Scoped, least privilege access to Firecrawl resources
- Full audit trail of agent actions to support review and compliance

## Use Firecrawl with any AI Agent Framework

Choose a framework you want to connect Firecrawl with:

- [ChatGPT](https://composio.dev/toolkits/firecrawl/framework/chatgpt)
- [OpenAI Agents SDK](https://composio.dev/toolkits/firecrawl/framework/open-ai-agents-sdk)
- [Claude Agent SDK](https://composio.dev/toolkits/firecrawl/framework/claude-agents-sdk)
- [Claude Code](https://composio.dev/toolkits/firecrawl/framework/claude-code)
- [Claude Cowork](https://composio.dev/toolkits/firecrawl/framework/claude-cowork)
- [Codex](https://composio.dev/toolkits/firecrawl/framework/codex)
- [Cursor](https://composio.dev/toolkits/firecrawl/framework/cursor)
- [VS Code](https://composio.dev/toolkits/firecrawl/framework/vscode)
- [OpenCode](https://composio.dev/toolkits/firecrawl/framework/opencode)
- [OpenClaw](https://composio.dev/toolkits/firecrawl/framework/openclaw)
- [Hermes](https://composio.dev/toolkits/firecrawl/framework/hermes-agent)
- [Google ADK](https://composio.dev/toolkits/firecrawl/framework/google-adk)
- [LangChain](https://composio.dev/toolkits/firecrawl/framework/langchain)
- [Vercel AI SDK](https://composio.dev/toolkits/firecrawl/framework/ai-sdk)
- [Mastra AI](https://composio.dev/toolkits/firecrawl/framework/mastra-ai)
- [LlamaIndex](https://composio.dev/toolkits/firecrawl/framework/llama-index)
- [CrewAI](https://composio.dev/toolkits/firecrawl/framework/crew-ai)
- [Pydantic AI](https://composio.dev/toolkits/firecrawl/framework/pydantic-ai)
- [AutoGen](https://composio.dev/toolkits/firecrawl/framework/autogen)

## Related Toolkits

- [Tavily](https://composio.dev/toolkits/tavily) - Tavily offers powerful search and data retrieval from documents, databases, and the web. It helps teams locate and filter information instantly, saving hours on research.
- [Exa](https://composio.dev/toolkits/exa) - Exa is a data extraction and search platform for gathering and analyzing information from websites, APIs, or databases. It helps teams quickly surface insights and automate data-driven workflows.
- [Serpapi](https://composio.dev/toolkits/serpapi) - SerpApi is a real-time API for structured search engine results. It lets you automate SERP data collection, parsing, and analysis for SEO and research.
- [Peopledatalabs](https://composio.dev/toolkits/peopledatalabs) - Peopledatalabs delivers B2B data enrichment and identity resolution APIs. Supercharge your apps with accurate, up-to-date business and contact data.
- [Snowflake](https://composio.dev/toolkits/snowflake) - Snowflake is a cloud data warehouse built for elastic scaling, secure data sharing, and fast SQL analytics across major clouds.
- [Posthog](https://composio.dev/toolkits/posthog) - PostHog is an open-source analytics platform for tracking user interactions and product metrics. It helps teams refine features, analyze funnels, and reduce churn with actionable insights.
- [Amplitude](https://composio.dev/toolkits/amplitude) - Amplitude is a digital analytics platform for product and behavioral data insights. It helps teams analyze user journeys and make data-driven decisions quickly.
- [Bright Data MCP](https://composio.dev/toolkits/brightdata_mcp) - Bright Data MCP is an AI-powered web scraping and data collection platform. Instantly access public web data in real time with advanced scraping tools.
- [Browseai](https://composio.dev/toolkits/browseai) - Browseai is a web automation and data extraction platform that turns any website into an API. It's perfect for monitoring websites and retrieving structured data without manual scraping.
- [ClickHouse](https://composio.dev/toolkits/clickhouse) - ClickHouse is an open-source, column-oriented database for real-time analytics and big data processing using SQL. Its lightning-fast query performance makes it ideal for handling large datasets and delivering instant insights.
- [Coinmarketcal](https://composio.dev/toolkits/coinmarketcal) - CoinMarketCal is a community-powered crypto calendar for upcoming events, announcements, and releases. It helps traders track market-moving developments and stay ahead in the crypto space.
- [Control d](https://composio.dev/toolkits/control_d) - Control d is a customizable DNS filtering and traffic redirection platform. It helps you manage internet access, enforce policies, and monitor usage across devices and networks.
- [Databox](https://composio.dev/toolkits/databox) - Databox is a business analytics platform that connects your data from any tool and device. It helps you track KPIs, build dashboards, and discover actionable insights.
- [Databricks](https://composio.dev/toolkits/databricks) - Databricks is a unified analytics platform for big data and AI on the lakehouse architecture. It empowers data teams to collaborate, analyze, and build scalable solutions efficiently.
- [Datagma](https://composio.dev/toolkits/datagma) - Datagma delivers data intelligence and analytics for business growth and market discovery. Get actionable market insights and track competitors to inform your strategy.
- [Delighted](https://composio.dev/toolkits/delighted) - Delighted is a customer feedback platform based on the Net Promoter System®. It helps you quickly gather, track, and act on customer sentiment.
- [Dovetail](https://composio.dev/toolkits/dovetail) - Dovetail is a research analysis platform for transcript review and insight generation. It helps teams code interviews, analyze feedback, and create actionable research summaries.
- [Dub](https://composio.dev/toolkits/dub) - Dub is a short link management platform with analytics and API access. Use it to easily create, manage, and track branded short links for your business.
- [Elasticsearch](https://composio.dev/toolkits/elasticsearch) - Elasticsearch is a distributed, RESTful search and analytics engine for all types of data. It delivers fast, scalable search and powerful analytics across massive datasets.
- [Fireflies](https://composio.dev/toolkits/fireflies) - Fireflies.ai is an AI-powered meeting assistant that records, transcribes, and analyzes voice conversations. It helps teams capture call notes automatically and search or summarize meetings effortlessly.

## Frequently Asked Questions

### Do I need my own developer credentials to use Firecrawl with Composio?

Yes, Firecrawl requires you to configure your own API key credentials. Once set up, Composio handles secure credential storage and API request handling for you.

### Can I use multiple toolkits together?

Yes! Composio's Tool Router enables agents to use multiple toolkits. [Learn more](https://docs.composio.dev/tool-router/overview).

### Is Composio secure?

Composio is SOC 2 and ISO 27001 compliant with all data encrypted in transit and at rest. [Learn more](https://trust.composio.dev).

### What if the API changes?

Composio maintains and updates all toolkit integrations automatically, so your agents always work with the latest API versions.

---
[See all toolkits](https://composio.dev/toolkits) · [Composio docs](https://docs.composio.dev/llms.txt)
