How to integrate Firecrawl MCP with Codex

Trusted by
AWS
Glean
Zoom
Airtable

30 min · no commitment · see it on your stack

Firecrawl logo
Codex logo
divider

Introduction

Codex is one of the most popular coding harnesses out there. And MCP makes the experience even better. With Firecrawl MCP integration, you can draft, triage, summarise emails, and much more, all without leaving the terminal or the app, whichever you prefer.

Also integrate Firecrawl with

Why use Composio?

Apart from a managed and hosted MCP server, you will get:

  • CodeAct: A dedicated workbench that allows GPT to write its code to handle complex tool chaining. Reduces to-and-fro with LLMs for frequent tool calling.
  • Large tool responses: Handle them to minimise context rot.
  • Dynamic just-in-time access to 20,000 tools across 1000+ other Apps for cross-app workflows. It loads the tools you need, so GPTs aren't overwhelmed by tools you don't need.

How to install Firecrawl MCP in Codex

Run the setup command

Run this command in your terminal to add the Composio MCP server to Codex.

Terminal

It will initiate the authentication in a browser window, authorize Codex to access your Composio account.

Composio authentication page

(Optional) Authenticate with OAuth

To authenticate manually, run the login command to open a browser window and authorize Codex to access your Composio account.

bash
codex mcp login composio

Verify the connection

Run codex mcp list to confirm Composio appears as a registered MCP server.

bash
codex mcp list

Codex App

Codex App follows the same approach as VS Code.

  1. Click ⚙️ on the bottom left → MCP Servers → + Add servers → Streamable HTTP:
  2. Fill the header and Key fields with { "x-consumer-api-key" = "ck_*******" }.
  3. The Key is the Composio API key, that you can find on dashboard.composio.dev
  4. Click on Authenticate and authorize Codex to your Composio account and you're all set.
Codex App MCP setup
  1. Restart and verify if it's there in .codex/config.toml
bash
[mcp_servers.composio]
url = "https://connect.composio.dev/mcp"
http_headers = { "x-consumer-api-key" = "ck_*******" }

What is the Firecrawl MCP server, and what's possible with it?

The Firecrawl MCP server is an implementation of the Model Context Protocol that connects your AI agent and assistants like Claude, Cursor, etc directly to your Firecrawl account. It provides structured and secure access to automated web crawling, scraping, and data extraction, so your agent can perform actions like indexing sites, extracting structured content, mapping URLs, and searching the web on your behalf.

  • Automated web crawling and indexing: Let your agent launch and manage web crawl jobs to gather content or index entire websites efficiently.
  • Structured data extraction: Instruct your agent to extract targeted data from web pages using custom prompts or schemas, turning unstructured sites into actionable information.
  • URL mapping and discovery: Have the agent explore and map all URLs within a website, including options for subdomain inclusion, sitemap processing, or search-based discovery.
  • On-demand scraping and content retrieval: Enable your agent to scrape specific URLs, retrieve page content, and even extract structured JSON using LLM-powered methods.
  • Integrated web search and data collection: Task your agent with running web searches, scraping top result pages, and returning relevant details—all in one workflow.

Supported Tools & Triggers

Tools
Cancel an agent jobTool to cancel an in-progress agent job by its ID.
Batch scrape multiple URLsTool to scrape multiple URLs in batch with concurrent processing.
Cancel a batch scrape jobTool to cancel a running batch scrape job using its unique identifier.
Get batch scrape statusRetrieves the current status and results of a batch scrape job using the job ID.
Get errors from batch scrape jobTool to retrieve error details from a batch scrape job, including failed URLs and URLs blocked by robots.
Start a web crawlInitiates a Firecrawl web crawl from a given URL, applying various filtering and content extraction rules, and polls until the job is complete; ensure the URL is accessible and any regex patterns for paths are valid.
Cancel a crawl jobCancels an active or queued web crawl job using its ID; attempting to cancel completed, failed, or previously canceled jobs will not change their state.
Cancel a crawl jobTool to cancel a running crawl job by its ID.
Get crawl job statusTool to retrieve the status and results of a Firecrawl crawl job.
Get errors from a crawl jobTool to retrieve errors from a Firecrawl crawl job.
Get all active crawl jobsTool to retrieve all active crawl jobs for the authenticated team.
Preview crawl parametersPreview crawl parameters before starting a crawl by generating optimal configuration from natural language instructions.
Start a web crawl (v2) [NEW][NEW v2 API] Initiates a Firecrawl v2 web crawl with enhanced features over v1: natural language prompts for automatic crawler configuration, crawlEntireDomain for sibling/parent page discovery, better depth control with maxDiscoveryDepth, subdomain support, and full webhook configuration.
Get team credit usageTool to get current team credit usage information.
Get historical team credit usageTool to retrieve historical team credit usage on a monthly basis.
Extract structured dataExtracts structured data from web pages by initiating an extraction job and polling for completion; requires a natural language `prompt` or a JSON `schema` (one must be provided).
Get extract job statusTool to retrieve the status and results of a previously submitted extract job.
Get agent job statusTool to get the status and results of an agent job.
Get deep research statusRetrieves the status and results of a deep research job by its ID.
Get the status of a crawl jobRetrieves the current status, progress, and details of a web crawl job, using the job ID obtained when the crawl was initiated.
Generate LLMs.txt for a websiteInitiates an async job to generate an LLMs.
Get LLMs.txt generation job statusTool to get the status and results of an LLMs.
Map multiple URLsMaps a website by discovering URLs from a starting base URL, with options to customize the crawl via search query, subdomain inclusion, sitemap handling, and result limits; search effectiveness is site-dependent.
Get team queue statusTool to retrieve metrics about the team's scrape queue.
Scrape URLScrapes a publicly accessible URL, optionally performing pre-scrape browser actions or extracting structured JSON using an LLM, to retrieve content in specified formats.
SearchPerforms a web search for a query, scrapes content from the top search results using Firecrawl, and returns details in specified formats.
Start an agent jobTool to start an agent job for agentic web extraction with multi-page navigation and interaction capabilities.
Get team token usageTool to retrieve the current team's token usage and balance information for Firecrawl's Extract feature.
Get historical team token usageTool to retrieve historical team token usage on a monthly basis.

Conclusion

You've successfully integrated Firecrawl with Codex using Composio's MCP server. Now you can interact with Firecrawl directly from your terminal, VS Code, or the Codex App using natural language commands.

Key benefits of this setup:

  • Seamless integration across CLI, VS Code, and standalone app
  • Natural language commands for Firecrawl operations
  • Managed authentication through Composio
  • Access to 20,000+ tools across 1000+ apps for cross-app workflows
  • CodeAct workbench for complex tool chaining

Next steps:

  • Try asking Codex to perform various Firecrawl operations
  • Explore cross-app workflows by connecting more toolkits
  • Build automation scripts that leverage Codex's AI capabilities

How to build Firecrawl MCP Agent with another framework

FAQ

What are the differences in Tool Router MCP and Firecrawl MCP?

With a standalone Firecrawl MCP server, the agents and LLMs can only access a fixed set of Firecrawl tools tied to that server. However, with the Composio Tool Router, agents can dynamically load tools from Firecrawl and many other apps based on the task at hand, all through a single MCP endpoint.

Can I use Tool Router MCP with Codex?

Yes, you can. Codex fully supports MCP integration. You get structured tool calling, message history handling, and model orchestration while Tool Router takes care of discovering and serving the right Firecrawl tools.

Can I manage the permissions and scopes for Firecrawl while using Tool Router?

Yes, absolutely. You can configure which Firecrawl scopes and actions are allowed when connecting your account to Composio. You can also bring your own OAuth credentials or API configuration so you keep full control over what the agent can do.

How safe is my data with Composio Tool Router?

All sensitive data such as tokens, keys, and configuration is fully encrypted at rest and in transit. Composio is SOC 2 Type 2 compliant and follows strict security practices so your Firecrawl data and credentials are handled as safely as possible.

Used by agents from

Context
Letta
glean
HubSpot
Agent.ai
Altera
DataStax
Entelligence
Rolai
Context
Letta
glean
HubSpot
Agent.ai
Altera
DataStax
Entelligence
Rolai
Context
Letta
glean
HubSpot
Agent.ai
Altera
DataStax
Entelligence
Rolai

Never worry about agent reliability

We handle tool reliability, observability, and security so you never have to second-guess an agent action.