# How to integrate Kaggle MCP with Google ADK

```json
{
  "title": "How to integrate Kaggle MCP with Google ADK",
  "toolkit": "Kaggle",
  "toolkit_slug": "kaggle",
  "framework": "Google ADK",
  "framework_slug": "google-adk",
  "url": "https://composio.dev/toolkits/kaggle/framework/google-adk",
  "markdown_url": "https://composio.dev/toolkits/kaggle/framework/google-adk.md",
  "updated_at": "2026-05-12T10:16:40.031Z"
}
```

## Introduction

This guide walks you through connecting Kaggle to Google ADK using the Composio tool router. By the end, you'll have a working Kaggle agent that can download data files for the titanic competition, create a new version of your covid-19 dataset, check processing status of your uploaded dataset through natural language commands.
This guide will help you understand how to give your Google ADK agent real control over a Kaggle account through Composio's Kaggle MCP server.
Before we dive in, let's take a quick look at the key ideas and tools involved.

## Also integrate Kaggle with

- [ChatGPT](https://composio.dev/toolkits/kaggle/framework/chatgpt)
- [OpenAI Agents SDK](https://composio.dev/toolkits/kaggle/framework/open-ai-agents-sdk)
- [Claude Agent SDK](https://composio.dev/toolkits/kaggle/framework/claude-agents-sdk)
- [Claude Code](https://composio.dev/toolkits/kaggle/framework/claude-code)
- [Claude Cowork](https://composio.dev/toolkits/kaggle/framework/claude-cowork)
- [Codex](https://composio.dev/toolkits/kaggle/framework/codex)
- [Cursor](https://composio.dev/toolkits/kaggle/framework/cursor)
- [VS Code](https://composio.dev/toolkits/kaggle/framework/vscode)
- [OpenCode](https://composio.dev/toolkits/kaggle/framework/opencode)
- [OpenClaw](https://composio.dev/toolkits/kaggle/framework/openclaw)
- [Hermes](https://composio.dev/toolkits/kaggle/framework/hermes-agent)
- [CLI](https://composio.dev/toolkits/kaggle/framework/cli)
- [LangChain](https://composio.dev/toolkits/kaggle/framework/langchain)
- [Vercel AI SDK](https://composio.dev/toolkits/kaggle/framework/ai-sdk)
- [Mastra AI](https://composio.dev/toolkits/kaggle/framework/mastra-ai)
- [LlamaIndex](https://composio.dev/toolkits/kaggle/framework/llama-index)
- [CrewAI](https://composio.dev/toolkits/kaggle/framework/crew-ai)

## TL;DR

Here's what you'll learn:
- Get a Kaggle account set up and connected to Composio
- Install the Google ADK and Composio packages
- Create a Composio Tool Router session for Kaggle
- Build an agent that connects to Kaggle through MCP
- Interact with Kaggle using natural language

## What is Google ADK?

Google ADK (Agents Development Kit) is Google's framework for building AI agents powered by Gemini models. It provides tools for creating agents that can use external services through the Model Context Protocol.
Key features include:
- Gemini Integration: Native support for Google's Gemini models
- MCP Toolset: Built-in support for Model Context Protocol tools
- Streamable HTTP: Connect to external services through streamable HTTP
- CLI and Web UI: Run agents via command line or web interface

## What is the Kaggle MCP server, and what's possible with it?

The Kaggle MCP server is an implementation of the Model Context Protocol that connects your AI agent and assistants like Claude, Cursor, etc directly to your Kaggle account. It provides structured and secure access to your Kaggle datasets, competitions, and configurations, so your agent can perform actions like downloading competition data, creating datasets, submitting entries, and managing dataset versions on your behalf.
- Competition data access and download: Let your agent fetch and download competition datasets quickly by specifying a competition ID, so you always have the latest files for analysis.
- Automated dataset creation and management: Have your agent create new Kaggle datasets, update metadata, and publish new dataset versions seamlessly, streamlining the process of sharing your work with the community.
- Competition entry submission: Empower your agent to submit competition entries automatically once your solution is ready and uploaded, helping you participate in challenges without manual hassle.
- Configuration management and setup: Allow your agent to initialize, locate, and update Kaggle API configuration files and keys, ensuring smooth and authenticated operations every time.
- Dataset status monitoring: Ask your agent to check the status of uploaded datasets or processing jobs, so you always know when your data is ready for use or public sharing.

## Supported Tools

| Tool slug | Name | Description |
|---|---|---|
| `KAGGLE_COMPETITION_DOWNLOAD_FILES` | Download competition data files | Downloads all data files for a Kaggle competition as a single zip archive. Returns the local file path where the zip was saved. Note: You must have accepted the competition's rules on Kaggle's website before downloading (403 error if not accepted). |
| `KAGGLE_COMPETITION_SUBMIT` | Submit Competition Entry | Submit an entry to a Kaggle competition using a previously uploaded file. Prerequisites: 1. You must have accepted the competition rules on Kaggle's website 2. You must have uploaded your submission file and obtained a blob_file_tokens (use Kaggle's file upload API endpoint first) This action performs the final submission step after file upload. The blob token identifies your uploaded file and associates it with your competition submission. |
| `KAGGLE_CONFIG_DIR` | Get Kaggle Config Directory | Tool to retrieve the directory of the Kaggle API configuration file. Use when you need to locate the directory containing your kaggle.json credentials. |
| `KAGGLE_CONFIG_INIT` | Initialize Kaggle Configuration | Initialize Kaggle API client configuration. This action sets up the necessary configuration file for Kaggle API access by first attempting to use the Kaggle CLI's 'kaggle config init' command. If the CLI is unavailable, it falls back to creating a kaggle.json file at ~/.kaggle/kaggle.json (or $KAGGLE_CONFIG_DIR/kaggle.json if that environment variable is set). The action is idempotent - if configuration already exists, it will not overwrite it. No parameters are required; the action uses environment variables and metadata when available. Run this before other Kaggle actions when credentials are missing or when KAGGLE_CONFIG_VIEW returns empty/error output. |
| `KAGGLE_CONFIG_KEYS` | List Kaggle Configuration Keys | Tool to list local Kaggle API configuration keys. Use when you need to see which configuration options are set without revealing values. |
| `KAGGLE_CONFIG_PATH` | Get Kaggle Config Path | Tool to retrieve local Kaggle API configuration file path. Use when you need to know the location of the Kaggle config before operations. |
| `KAGGLE_CONFIG_RESET` | Reset Kaggle Configuration | Tool to reset local Kaggle CLI configuration to defaults. Clears CLI-managed keys ('competition', 'path', 'proxy'). |
| `KAGGLE_CONFIG_SET` | Set Kaggle Configuration | Tool to set a Kaggle CLI configuration parameter. Use when updating local CLI settings such as default download path or proxy. Ensure Kaggle CLI is installed. |
| `KAGGLE_CONFIG_UNSET` | Unset Kaggle Configuration | Tool to unset a Kaggle CLI configuration parameter. Use when removing local CLI settings such as default download path or proxy. Ensure Kaggle CLI is installed. |
| `KAGGLE_CONFIG_VIEW` | View Kaggle Configuration | View local Kaggle API credentials and configuration settings. This action reads Kaggle configuration from local sources (does NOT make API calls to Kaggle). Configuration is retrieved in the following precedence order: 1. kaggle.json file (from KAGGLE_CONFIG_DIR env var, ~/.config/kaggle/, or ~/.kaggle/) 2. 'kaggle config view' CLI output (for proxy/path settings) 3. Environment variables (KAGGLE_USERNAME, KAGGLE_KEY) 4. Authorization header from metadata Use this action to: - Verify Kaggle credentials are configured before making API calls - Check current proxy settings - Debug authentication issues Returns empty strings for username/key if no credentials are found; use KAGGLE_CONFIG_INIT to set up credentials first. Note: username and key are independent — an empty username field does not indicate missing or invalid credentials. WARNING: This action returns sensitive API key data in plain text. |
| `KAGGLE_DATASET_CREATE` | Dataset Create | Create a new Kaggle dataset with metadata. IMPORTANT: Dataset creation requires at least one data file. Ensure files are uploaded before calling this action. The 'id' field must use your authenticated Kaggle username as the owner. Returns the creation status and any message from the Kaggle API. |
| `KAGGLE_DATASET_INIT` | Kaggle Dataset Init | Tool to initialize a dataset-metadata.json file in a local folder. Use when preparing a dataset folder before uploading to Kaggle. |
| `KAGGLE_DATASET_LIST_FILES` | List Kaggle Dataset Files | Tool to list files in a Kaggle dataset. Use when you need to retrieve paginated file listings by owner and dataset slugs, with optional version and paging controls. |
| `KAGGLE_DATASET_STATUS` | Get Dataset Status | Check the processing status of a Kaggle dataset after creation or version update. This endpoint is used to monitor datasets that are currently being processed by Kaggle's servers. It returns status information for datasets that are actively uploading, processing, or experiencing errors. For already-published datasets, this endpoint typically returns 404 (Not Found), which is expected behavior. Use this tool immediately after creating a new dataset (KAGGLE_DATASET_CREATE) or updating an existing dataset version (KAGGLE_DATASET_VERSION) to check when the dataset becomes ready. Poll this endpoint periodically until the status indicates completion or error. |
| `KAGGLE_DATASET_VERSION` | Create Dataset Version | Create a new version of an existing Kaggle dataset. Prerequisites: - You must own the dataset or have edit permissions - Files must be uploaded first to obtain upload tokens (required for the 'files' parameter) Use this when you have updated files or metadata and need to publish a new version of an existing dataset. |
| `KAGGLE_DOWNLOAD_COMPETITION_FILE` | Download competition file | Tool to download a specific data file from a Kaggle competition. Use when you need to retrieve a single file from a competition by specifying the competition slug and filename. Note: You must have accepted the competition's rules on Kaggle's website before downloading. |
| `KAGGLE_DOWNLOAD_COMPETITION_LEADERBOARD` | Download competition leaderboard | Tool to download the entire competition leaderboard as a CSV file packaged in a ZIP archive. Use when you need to analyze or review competition standings and scores. |
| `KAGGLE_DOWNLOAD_DATASET` | Download Kaggle Dataset | Tool to download all files from a Kaggle dataset as a zip archive. Supports downloading specific versions by providing the dataset_version_number parameter. |
| `KAGGLE_DOWNLOAD_DATASET_FILE` | Download Kaggle Dataset File | Tool to download a specific file from a Kaggle dataset. Use when you need to retrieve a single file from a dataset by specifying the owner, dataset, and filename. |
| `KAGGLE_GENERATE_COMPETITION_SUBMISSION_URL` | Generate Competition Submission URL | Tool to generate a pre-signed URL for uploading competition submission files. Use this before uploading your submission file to Kaggle. This action generates a temporary upload URL and token for submitting to a competition. You must provide the competition ID, file size, and last modified timestamp. After obtaining the URL, upload your submission file to the createUrl, then use the token to finalize the submission. |
| `KAGGLE_GET_DATASET_METADATA` | Get Dataset Metadata | Tool to get comprehensive metadata for a Kaggle dataset including title, description, licenses, and tags. Use when you need detailed information about a dataset's structure, schema, or properties. |
| `KAGGLE_GET_MODEL` | Get Model Details | Tool to get a Kaggle model's details including metadata and description. Use when you need information about a specific model on Kaggle. |
| `KAGGLE_GET_MODEL_INSTANCE` | Get Model Instance Details | Tool to get details for a specific Kaggle model instance (variation). Returns metadata including overview, usage instructions, download URL, version information, and license details. Use when you need to inspect or retrieve information about a specific model variation before downloading or using it. |
| `KAGGLE_KERNEL_INIT` | Kaggle Kernel Init | Initialize a kernel-metadata.json template file in a specified folder. This file is required before pushing/uploading a kernel to Kaggle. The template includes default values for kernel configuration (language, kernel_type, GPU settings, etc.) that can be customized before pushing. Use this when setting up a new Kaggle kernel locally. |
| `KAGGLE_KERNEL_OUTPUT` | Download kernel output | Tool to download the output of a Kaggle kernel. Use when needing the latest kernel results locally. |
| `KAGGLE_KERNELS_STATUS` | Get Kernel Status | Get the execution status of a Kaggle kernel (notebook). Returns current status (running, complete, error), timestamps, and output URL. Use this to monitor kernel execution after pushing/submitting a kernel. Note: You need permission to access the kernel - typically only your own kernels or public kernels you have access to. |
| `KAGGLE_LIST_COMPETITION_FILES` | List competition data files | Tool to list all data files available for a Kaggle competition. Use when you need to retrieve file names, sizes, and metadata for competition datasets before downloading. |
| `KAGGLE_LIST_COMPETITIONS` | List Kaggle Competitions | Tool to list available Kaggle competitions with filters and pagination. Use when you need to discover competitions, search by keywords, or filter by category, group, and sorting options. |
| `KAGGLE_LIST_DATASETS` | List Kaggle Datasets | Tool to list Kaggle datasets with filters and pagination. Use after authenticating with Kaggle API key. |
| `KAGGLE_LIST_KERNEL_OUTPUT_FILES` | List Kernel Output Files | Tool to list output files for a specific kernel run. Use when you need to retrieve paginated file listings by kernel owner and slug. |
| `KAGGLE_LIST_KERNELS` | List Kaggle Kernels | Tool to list Kaggle kernels (notebooks and scripts) with filters and pagination. Use to discover kernels by search terms, user, language, type, competition, or dataset. |
| `KAGGLE_LIST_MODEL_INSTANCE_VERSION_FILES` | List Model Instance Version Files | Tool to list files for a specific version of a model variation. Use when you need to retrieve files for a particular model framework instance version by owner, model, framework, variation, and version. |
| `KAGGLE_LIST_MODELS` | List Kaggle Models | Tool to list Kaggle models with optional filters for owner, sorting, search, and pagination. Use to discover available models on Kaggle's platform. |
| `KAGGLE_PULL_KERNEL` | Pull Kernel Code | Tool to pull (download) the source code of a Kaggle kernel to local storage. Use when you need to retrieve a kernel's notebook, script, or metadata files. Optionally include metadata JSON file with kernel configuration details. |
| `KAGGLE_VIEW_COMPETITION_LEADERBOARD` | View competition leaderboard | Tool to view competition leaderboard information showing rankings and scores of participants. Use when you need to check competition standings, team scores, or analyze leaderboard positions. |

## Supported Triggers

None listed.

## Creating MCP Server - Stand-alone vs Composio SDK

The Kaggle MCP server is an implementation of the Model Context Protocol that connects your AI agent to Kaggle. It provides structured and secure access so your agent can perform Kaggle operations on your behalf through a secure, permission-based interface.
With Composio's managed implementation, you don't have to create your own developer app. For production, if you're building an end product, we recommend using your own credentials. The managed server helps you prototype fast and go from 0-1 faster.

## Step-by-step Guide

### 1. Prerequisites

Before starting, make sure you have:
- A Google API key for Gemini models
- A Composio account and API key
- Python 3.9 or later installed
- Basic familiarity with Python

### 1. Getting API Keys for Google and Composio

Google API Key
- Go to [Google AI Studio](https://aistudio.google.com/app/apikey) and create an API key.
- Copy the key and keep it safe. You will put this in GOOGLE_API_KEY.
Composio API Key and User ID
- Log in to the [Composio dashboard](https://dashboard.composio.dev?utm_source=toolkits&utm_medium=framework_docs).
- Go to Settings → API Keys and copy your Composio API key. Use this for COMPOSIO_API_KEY.
- Decide on a stable user identifier to scope sessions, often your email or a user ID. Use this for COMPOSIO_USER_ID.

### 2. Install dependencies

Inside your virtual environment, install the required packages.
What's happening:
- google-adk is Google's Agents Development Kit
- composio connects your agent to Kaggle via MCP
- python-dotenv loads environment variables
```bash
pip install google-adk composio python-dotenv
```

### 3. Set up ADK project

Set up a new Google ADK project.
What's happening:
- This creates an agent folder with a root agent file and .env file
```bash
adk create my_agent
```

### 4. Set environment variables

Save all your credentials in the .env file.
What's happening:
- GOOGLE_API_KEY authenticates with Google's Gemini models
- COMPOSIO_API_KEY authenticates with Composio
- COMPOSIO_USER_ID identifies the user for session management
```bash
GOOGLE_API_KEY=your-google-api-key
COMPOSIO_API_KEY=your-composio-api-key
COMPOSIO_USER_ID=your-user-id-or-email
```

### 5. Import modules and validate environment

What's happening:
- os reads environment variables
- Composio is the main Composio SDK client
- GoogleProvider declares that you are using Google ADK as the agent runtime
- Agent is the Google ADK LLM agent class
- McpToolset lets the ADK agent call MCP tools over HTTP
```python
import os
import warnings

from composio import Composio
from dotenv import load_dotenv
from google.adk.agents.llm_agent import Agent
from google.adk.tools.mcp_tool.mcp_session_manager import StreamableHTTPConnectionParams
from google.adk.tools.mcp_tool.mcp_toolset import McpToolset

load_dotenv()

warnings.filterwarnings("ignore", message=".*BaseAuthenticatedTool.*")

GOOGLE_API_KEY = os.getenv("GOOGLE_API_KEY")
COMPOSIO_API_KEY = os.getenv("COMPOSIO_API_KEY")
COMPOSIO_USER_ID = os.getenv("COMPOSIO_USER_ID")

if not GOOGLE_API_KEY:
    raise ValueError("GOOGLE_API_KEY is not set in the environment.")
if not COMPOSIO_API_KEY:
    raise ValueError("COMPOSIO_API_KEY is not set in the environment.")
if not COMPOSIO_USER_ID:
    raise ValueError("COMPOSIO_USER_ID is not set in the environment.")
```

### 6. Create Composio client and Tool Router session

What's happening:
- Authenticates to Composio with your API key
- Declares Google ADK as the provider
- Spins up a short-lived MCP endpoint for your user and selected toolkit
- Stores the MCP HTTP URL for the ADK MCP integration
```python
composio_client = Composio(api_key=COMPOSIO_API_KEY)

composio_session = composio_client.create(
    user_id=COMPOSIO_USER_ID,
    toolkits=["kaggle"],
)

COMPOSIO_MCP_URL = composio_session.mcp.url,
print(f"Composio MCP URL: {COMPOSIO_MCP_URL}")
```

### 7. Set up the McpToolset and create the Agent

What's happening:
- Connects the ADK agent to the Composio MCP endpoint through McpToolset
- Uses Gemini as the model powering the agent
- Lists exact tool names in instruction to reduce misnamed tool calls
```python
composio_toolset = McpToolset(
    connection_params=StreamableHTTPConnectionParams(
        url=COMPOSIO_MCP_URL,
        headers={"x-api-key": COMPOSIO_API_KEY}
    )
)

root_agent = Agent(
    model="gemini-2.5-flash",
    name="composio_agent",
    description="An agent that uses Composio tools to perform actions.",
    instruction=(
        "You are a helpful assistant connected to Composio. "
        "You have the following tools available: "
        "COMPOSIO_SEARCH_TOOLS, COMPOSIO_MULTI_EXECUTE_TOOL, "
        "COMPOSIO_MANAGE_CONNECTIONS, COMPOSIO_REMOTE_BASH_TOOL, COMPOSIO_REMOTE_WORKBENCH. "
        "Use these tools to help users with Kaggle operations."
    ),
    tools=[composio_toolset],
)

print("\nAgent setup complete. You can now run this agent directly ;)")
```

### 8. Run the agent

Execute the agent from the project root. The web command opens a web portal where you can chat with the agent.
What's happening:
- adk run runs the agent in CLI mode
- adk web . opens a web UI for interactive testing
```bash
# Run in CLI mode
adk run my_agent

# Or run in web UI mode
adk web
```

## Complete Code

```python
import os
import warnings

from composio import Composio
from composio_google import GoogleProvider
from dotenv import load_dotenv
from google.adk.agents.llm_agent import Agent
from google.adk.tools.mcp_tool.mcp_session_manager import StreamableHTTPConnectionParams
from google.adk.tools.mcp_tool.mcp_toolset import McpToolset

load_dotenv()
warnings.filterwarnings("ignore", message=".*BaseAuthenticatedTool.*")

GOOGLE_API_KEY = os.getenv("GOOGLE_API_KEY")
COMPOSIO_API_KEY = os.getenv("COMPOSIO_API_KEY")
COMPOSIO_USER_ID = os.getenv("COMPOSIO_USER_ID")

if not GOOGLE_API_KEY:
    raise ValueError("GOOGLE_API_KEY is not set in the environment.")
if not COMPOSIO_API_KEY:
    raise ValueError("COMPOSIO_API_KEY is not set in the environment.")
if not COMPOSIO_USER_ID:
    raise ValueError("COMPOSIO_USER_ID is not set in the environment.")

composio_client = Composio(api_key=COMPOSIO_API_KEY, provider=GoogleProvider())

composio_session = composio_client.create(
    user_id=COMPOSIO_USER_ID,
    toolkits=["kaggle"],
)

COMPOSIO_MCP_URL = composio_session.mcp.url


composio_toolset = McpToolset(
    connection_params=StreamableHTTPConnectionParams(
        url=COMPOSIO_MCP_URL,
        headers={"x-api-key": COMPOSIO_API_KEY}
    )
)

root_agent = Agent(
    model="gemini-2.5-flash",
    name="composio_agent",
    description="An agent that uses Composio tools to perform actions.",
    instruction=(
        "You are a helpful assistant connected to Composio. "
        "You have the following tools available: "
        "COMPOSIO_SEARCH_TOOLS, COMPOSIO_MULTI_EXECUTE_TOOL, "
        "COMPOSIO_MANAGE_CONNECTIONS, COMPOSIO_REMOTE_BASH_TOOL, COMPOSIO_REMOTE_WORKBENCH. "
        "Use these tools to help users with Kaggle operations."
    ),  
    tools=[composio_toolset],
)

print("\nAgent setup complete. You can now run this agent directly ;)")
```

## Conclusion

You've successfully integrated Kaggle with the Google ADK through Composio's MCP Tool Router. Your agent can now interact with Kaggle using natural language commands.
Key takeaways:
- The Tool Router approach dynamically routes requests to the appropriate Kaggle tools
- Environment variables keep your credentials secure and separate from code
- Clear agent instructions reduce tool calling errors
- The ADK web UI provides an interactive interface for testing and development
You can extend this setup by adding more toolkits to the toolkits array in your session configuration.

## How to build Kaggle MCP Agent with another framework

- [ChatGPT](https://composio.dev/toolkits/kaggle/framework/chatgpt)
- [OpenAI Agents SDK](https://composio.dev/toolkits/kaggle/framework/open-ai-agents-sdk)
- [Claude Agent SDK](https://composio.dev/toolkits/kaggle/framework/claude-agents-sdk)
- [Claude Code](https://composio.dev/toolkits/kaggle/framework/claude-code)
- [Claude Cowork](https://composio.dev/toolkits/kaggle/framework/claude-cowork)
- [Codex](https://composio.dev/toolkits/kaggle/framework/codex)
- [Cursor](https://composio.dev/toolkits/kaggle/framework/cursor)
- [VS Code](https://composio.dev/toolkits/kaggle/framework/vscode)
- [OpenCode](https://composio.dev/toolkits/kaggle/framework/opencode)
- [OpenClaw](https://composio.dev/toolkits/kaggle/framework/openclaw)
- [Hermes](https://composio.dev/toolkits/kaggle/framework/hermes-agent)
- [CLI](https://composio.dev/toolkits/kaggle/framework/cli)
- [LangChain](https://composio.dev/toolkits/kaggle/framework/langchain)
- [Vercel AI SDK](https://composio.dev/toolkits/kaggle/framework/ai-sdk)
- [Mastra AI](https://composio.dev/toolkits/kaggle/framework/mastra-ai)
- [LlamaIndex](https://composio.dev/toolkits/kaggle/framework/llama-index)
- [CrewAI](https://composio.dev/toolkits/kaggle/framework/crew-ai)

## Related Toolkits

- [Composio](https://composio.dev/toolkits/composio) - Composio is an integration platform that connects AI agents with hundreds of business tools. It streamlines authentication and lets you trigger actions across services—no custom code needed.
- [Composio search](https://composio.dev/toolkits/composio_search) - Composio search is a unified web search toolkit spanning travel, e-commerce, news, financial markets, images, and more. It lets you and your apps tap into up-to-date web data from a single, easy-to-integrate service.
- [Perplexityai](https://composio.dev/toolkits/perplexityai) - Perplexityai delivers natural, conversational AI models for generating human-like text. Instantly get context-aware, high-quality responses for chat, search, or complex workflows.
- [Browser tool](https://composio.dev/toolkits/browser_tool) - Browser tool is a virtual browser integration that lets AI agents interact with the web programmatically. It enables automated browsing, scraping, and action-taking from any AI workflow.
- [Ai ml api](https://composio.dev/toolkits/ai_ml_api) - Ai ml api is a suite of AI/ML models for natural language and image tasks. It provides fast, scalable access to advanced AI capabilities for your apps and workflows.
- [Aivoov](https://composio.dev/toolkits/aivoov) - Aivoov is an AI-powered text-to-speech platform offering 1,000+ voices in over 150 languages. Instantly turn written content into natural, human-like audio for any application.
- [All images ai](https://composio.dev/toolkits/all_images_ai) - All-Images.ai is an AI-powered image generation and management platform. It helps you create, search, and organize images effortlessly with advanced AI capabilities.
- [Anthropic administrator](https://composio.dev/toolkits/anthropic_administrator) - Anthropic administrator is an API for managing Anthropic organizational resources like members, workspaces, and API keys. It helps you automate admin tasks and streamline resource management across your Anthropic organization.
- [Api labz](https://composio.dev/toolkits/api_labz) - Api labz is a platform offering a suite of AI-driven APIs and workflow tools. It helps developers automate tasks and build smarter, more efficient applications.
- [Apipie ai](https://composio.dev/toolkits/apipie_ai) - Apipie ai is an AI model aggregator offering a single API for accessing top AI models from multiple providers. It helps developers build cost-efficient, latency-optimized AI solutions without juggling multiple integrations.
- [Astica ai](https://composio.dev/toolkits/astica_ai) - Astica ai provides APIs for computer vision, NLP, and voice synthesis. Integrate advanced AI features into your app with a single API key.
- [Bigml](https://composio.dev/toolkits/bigml) - BigML is a machine learning platform that lets you build, train, and deploy predictive models from your data. Its intuitive interface and robust API make machine learning accessible and efficient.
- [Botbaba](https://composio.dev/toolkits/botbaba) - Botbaba is a platform for building, managing, and deploying conversational AI chatbots across messaging channels. It streamlines chatbot automation, making it easier to integrate AI into customer interactions.
- [Botpress](https://composio.dev/toolkits/botpress) - Botpress is an open-source platform for building, deploying, and managing chatbots. It helps teams automate conversations and deliver rich, interactive messaging experiences.
- [Chatbotkit](https://composio.dev/toolkits/chatbotkit) - Chatbotkit is a platform for building and managing AI-powered chatbots using robust APIs and SDKs. It lets you easily add conversational AI to your apps for better user engagement.
- [Cody](https://composio.dev/toolkits/cody) - Cody is an AI assistant built for businesses, trained on your company's knowledge and data. It delivers instant answers and insights, tailored for your team.
- [Context7 MCP](https://composio.dev/toolkits/context7_mcp) - Context7 MCP delivers live, version-specific code docs and examples right from the source. It helps developers and AI agents instantly retrieve authoritative programming info—no more out-of-date docs.
- [Customgpt](https://composio.dev/toolkits/customgpt) - CustomGPT.ai lets you build and deploy chatbots tailored to your own data and business needs. Get precise and context-aware AI conversations without writing code.
- [Datarobot](https://composio.dev/toolkits/datarobot) - Datarobot is a machine learning platform that automates model development, deployment, and monitoring. It empowers organizations to quickly gain predictive insights from large datasets.
- [Deepgram](https://composio.dev/toolkits/deepgram) - Deepgram is an AI-powered speech recognition platform for accurate audio transcription and understanding. It enables fast, scalable speech-to-text with advanced audio intelligence features.

## Frequently Asked Questions

### What are the differences in Tool Router MCP and Kaggle MCP?

With a standalone Kaggle MCP server, the agents and LLMs can only access a fixed set of Kaggle tools tied to that server. However, with the Composio Tool Router, agents can dynamically load tools from Kaggle and many other apps based on the task at hand, all through a single MCP endpoint.

### Can I use Tool Router MCP with Google ADK?

Yes, you can. Google ADK fully supports MCP integration. You get structured tool calling, message history handling, and model orchestration while Tool Router takes care of discovering and serving the right Kaggle tools.

### Can I manage the permissions and scopes for Kaggle while using Tool Router?

Yes, absolutely. You can configure which Kaggle scopes and actions are allowed when connecting your account to Composio. You can also bring your own OAuth credentials or API configuration so you keep full control over what the agent can do.

### How safe is my data with Composio Tool Router?

All sensitive data such as tokens, keys, and configuration is fully encrypted at rest and in transit. Composio is SOC 2 Type 2 compliant and follows strict security practices so your Kaggle data and credentials are handled as safely as possible.

---
[See all toolkits](https://composio.dev/toolkits) · [Composio docs](https://docs.composio.dev/llms.txt)
