Claude Context Mcp

@zilliz/claude-context-mcp

Model Context Protocol integration for Claude Context

0 downloads
v0.1.3

Capabilities

tools

Installation

Quick Install

Install using the MCPSearch CLI (recommended)

mcp install @zilliz/claude-context-mcp

Don't have the CLI? Install it first

Run with npx

Run directly without installing

npx -y @zilliz/claude-context-mcp

Manual Configuration

Add to your MCP client configuration file

CClaude Code / Claude Desktop

Add to ~/.claude/claude_desktop_config.json

{
  "mcpServers": {
    "claude-context-mcp": {
      "command": "npx",
      "args": [
        "-y",
        "@zilliz/claude-context-mcp"
      ]
    }
  }
}

CuCursor

Add to ~/.cursor/mcp.json

{
  "mcp": {
    "servers": {
      "claude-context-mcp": {
        "command": "npx",
        "args": [
          "-y",
          "@zilliz/claude-context-mcp"
        ]
      }
    }
  }
}

VSVS Code / Continue.dev

Add to .vscode/mcp.json or Continue settings

{
  "mcpServers": {
    "claude-context-mcp": {
      "command": "npx",
      "args": [
        "-y",
        "@zilliz/claude-context-mcp"
      ]
    }
  }
}

About

# @zilliz/claude-context-mcp ![](../../assets/claude-context.png) Model Context Protocol (MCP) integration for Claude Context - A powerful MCP server that enables AI assistants and agents to index and search codebases using semantic search. [![npm version](https://img.shields.io/npm/v/@zilliz/claude-context-mcp.svg)](https://www.npmjs.com/package/@zilliz/claude-context-mcp) [![npm downloads](https://img.shields.io/npm/dm/@zilliz/claude-context-mcp.svg)](https://www.npmjs.com/package/@zilliz/claude-context-mcp) > 📖 **New to Claude Context?** Check out the [main project README](../../README.md) for an overview and setup instructions. ## 🚀 Use Claude Context as MCP in Claude Code and others ![img](https://lh7-rt.googleusercontent.com/docsz/AD_4nXf2uIf2c5zowp-iOMOqsefHbY_EwNGiutkxtNXcZVJ8RI6SN9DsCcsc3amXIhOZx9VcKFJQLSAqM-2pjU9zoGs1r8GCTUL3JIsLpLUGAm1VQd5F2o5vpEajx2qrc77iXhBu1zWj?key=qYdFquJrLcfXCUndY-YRBQ) Model Context Protocol (MCP) allows you to integrate Claude Context with your favorite AI coding assistants, e.g. Claude Code. ## Quick Start ### Prerequisites Before using the MCP server, make sure you have: - API key for your chosen embedding provider (OpenAI, VoyageAI, Gemini, or Ollama setup) - Milvus vector database (local or cloud) > 💡 **Setup Help:** See the [main project setup guide](../../README.md#-quick-start) for detailed installation instructions. ### Prepare Environment Variables #### Embedding Provider Configuration Claude Context MCP supports multiple embedding providers. Choose the one that best fits your needs: > 📋 **Quick Reference**: For a complete list of environment variables and their descriptions, see the [Environment Variables Guide](../../docs/getting-started/environment-variables.md). ```bash # Supported providers: OpenAI, VoyageAI, Gemini, Ollama EMBEDDING_PROVIDER=OpenAI ``` <details> <summary><strong>1. OpenAI Configuration (Default)</strong></summary> OpenAI provides high-quality embeddings with excellent performance for code understanding. ```bash # Required: Your OpenAI API key OPENAI_API_KEY=sk-your-openai-api-key # Optional: Specify embedding model (default: text-embedding-3-small) EMBEDDING_MODEL=text-embedding-3-small # Optional: Custom API base URL (for Azure OpenAI or other compatible services) OPENAI_BASE_URL=https://api.openai.com/v1 ``` **Available Models:** See `getSupportedModels` in [`openai-embedding.ts`](https://github.com/zilliztech/claude-context/blob/master/packages/core/src/embedding/openai-embedding.ts) for the full list of supported models. **Getting API Key:** 1. Visit [OpenAI Platform](https://platform.openai.com/api-keys) 2. Sign in or create an account 3. Generate a new API key 4. Set up billing if needed </details> <details> <summary><strong>2. VoyageAI Configuration</strong></summary> VoyageAI offers specialized code embeddings optimized for programming languages. ```bash # Required: Your VoyageAI API key VOYAGEAI_API_KEY=pa-your-voyageai-api-key # Optional: Specify embedding model (default: voyage-code-3) EMBEDDING_MODEL=voyage-code-3 ``` **Available Models:** See `getSupportedModels` in [`voyageai-embedding.ts`](https://github.com/zilliztech/claude-context/blob/master/packages/core/src/embedding/voyageai-embedding.ts) for the full list of supported models. **Getting API Key:** 1. Visit [VoyageAI Console](https://dash.voyageai.com/) 2. Sign up for an account 3. Navigate to API Keys section 4. Create a new API key </details> <details> <summary><strong>3. Gemini Configuration</strong></summary> Google's Gemini provides competitive embeddings with good multilingual support. ```bash # Required: Your Gemini API key GEMINI_API_KEY=your-gemini-api-key # Optional: Specify embedding model (default: gemini-embedding-001) EMBEDDING_MODEL=gemini-embedding-001 ``` **Available Models:** See `getSupportedModels` in [`gemini-embedding.ts`](https://github.com/zilliztech/claude-context/blob/master/packages/core/src/embedding/gemini-embedding.ts) for the full list of supported models. **Getting API Key:** 1. Visit [Google AI Studio](https://aistudio.google.com/) 2. Sign in with your Google account 3. Go to "Get API key" section 4. Create a new API key </details> <details> <summary><strong>4. Ollama Configuration (Local/Self-hosted)</strong></summary> Ollama allows you to run embeddings locally without sending data to external services. ```bash # Required: Specify which Ollama model to use EMBEDDING_MODEL=nomic-embed-text # Optional: Specify Ollama host (default: http://127.0.0.1:11434) OLLAMA_HOST=http://127.0.0.1:11434 ``` **Setup Instructions:** 1. Install Ollama from [ollama.ai](https://ollama.ai/) 2. Pull the embedding model: ```bash ollama pull nomic-embed-text ``` 3. Ensure Ollama is running: ```bash ollama serve ``` </details> #### Get a free vector database on Zilliz Cloud Claude Context needs a vector database. You can [sign up](https://cloud.zilliz.com/signup?utm_source=github&utm_medium=referral&utm_campaign=2507-codecontext-readme) on Zilliz Cloud to get an API key. ![](../../assets/signup_and_get_apikey.png) Copy your Personal Key to replace `your-zilliz-cloud-api-key` in the configuration examples. ```bash MILVUS_TOKEN=your-zilliz-cloud-api-key ``` #### Embedding Batch Size You can set the embedding batch size to optimize the performance of the MCP server, depending on your embedding model throughput. The default value is 100. ```bash EMBEDDING_BATCH_SIZE=512 ``` #### Custom File Processing (Optional) You can configure custom file extensions and ignore patterns globally via environment variables: ```bash # Additional file extensions to include beyond defaults CUSTOM_EXTENSIONS=.vue,.svelte,.astro,.twig # Additional ignore patterns to exclude files/directories CUSTOM_IGNORE_PATTERNS=temp/**,*.backup,private/**,uploads/** ``` These settings work in combination with tool parameters - patterns from both sources will be merged together. ## Usage with MCP Clients <details> <summary><strong>Qwen Code</strong></summary> Create or edit the `~/.qwen/settings.json` file and add the following configuration: ```json { "mcpServers": { "claude-context": { "command": "npx", "args": ["@zilliz/claude-context-mcp@latest"], "env": { "OPENAI_API_KEY": "your-openai-api-key", "MILVUS_TOKEN": "your-zilliz-cloud-api-key" } } } } ``` </details> <details> <summary><strong>Cursor</strong></summary> Go to: `Settings` -> `Cursor Settings` -> `MCP` -> `Add new global MCP server` Pasting the following configuration into your Cursor `~/.cursor/mcp.json` file is the recommended approach. You may also install in a specific project by creating `.cursor/mcp.json` in your project folder. See [Cursor MCP docs](https://docs.cursor.com/context/model-context-protocol) for more info. **OpenAI Configuration (Default):** ```json { "mcpServers": { "claude-context": { "command": "npx", "args": ["-y", "@zilliz/claude-context-mcp@latest"], "env": { "EMBEDDING_PROVIDER": "OpenAI", "OPENAI_API_KEY": "your-openai-api-key", "MILVUS_TOKEN": "your-zilliz-cloud-api-key" } } } } ``` **VoyageAI Configuration:** ```json { "mcpServers": { "claude-context": { "command": "npx", "args": ["-y", "@zilliz/claude-context-mcp@latest"], "env": { "EMBEDDING_PROVIDER": "VoyageAI", "VOYAGEAI_API_KEY": "your-voyageai-api-key", "EMBEDDING_MODEL": "voyage-code-3", "MILVUS_TOKEN": "your-zilliz-cloud-api-key" } } } } ``` **Gemini Configuration:** ```json { "mcpServers": { "claude-context": { "command": "npx", "args": ["-y", "@zilliz/claude-context-mcp@latest"], "env": { "EMBEDDING_PROVIDER": "Gemini", "GEMINI_API_KEY": "your-gemini-api-key", "MILVUS_TOKEN": "your-zilliz-cloud-api-key" } } } } ``` **Ollama Configuration:** ```json { "mcpServers": { "claude-context": { "command": "npx", "args": ["-y", "@zilliz/claude-context-mcp@latest"], "env": { "EMBEDDING_PROVIDER": "Ollama", "EMBEDDING_MODEL": "nomic-embed-text", "OLLAMA_HOST": "http://127.0.0.1:11434", "MILVUS_TOKEN": "your-zilliz-cloud-api-key" } } } } ``` </details> <details> <summary><strong>Void</strong></summary> Go to: `Settings` -> `MCP` -> `Add MCP Server` Add the following configuration to your Void MCP settings: ```json { "mcpServers": { "code-context": { "command": "npx", "args": ["-y", "@zilliz/claude-context-mcp@latest"], "env": { "OPENAI_API_KEY": "your-openai-api-key", "MILVUS_ADDRESS": "your-zilliz-cloud-public-endpoint", "MILVUS_TOKEN": "your-zilliz-cloud-api-key" } } } } ``` </details> <details> <summary><strong>Claude Desktop</strong></summary> Add to your Claude Desktop configuration: ```json { "mcpServers": { "claude-context": { "command": "npx", "args": ["@zilliz/claude-context-mcp@latest"], "env": { "OPENAI_API_KEY": "your-openai-api-key", "MILVUS_TOKEN": "your-zilliz-cloud-api-key" } } } } ``` </details> <details> <summary><strong>Claude Code</strong></summary> Use the command line interface to add the Claude Context MCP server: ```bash # Add the Claude Context MCP server claude mcp add claude-context -e OPENAI_API_KEY=your-openai-api-key -e MILVUS_TOKEN=your-zilliz-cloud-api-key -- npx @zilliz/claude-context-mcp@latest ``` See the [Claude Code MCP documentation](https://docs.anthropic.com/en/docs/claude-code/mcp) for more details about MCP server management. </details> <details> <summary><strong>Windsurf</strong></summary> Windsurf supports MCP configuration through a JSON file. Add the following configuration to your Windsurf MCP settings: ```json { "mcpServers": { "claude-context": { "command": "npx", "args": ["-y", "@zilliz/claude-context-mcp@latest"], "env": { "OPENAI_API_KEY": "your-openai-api-key", "MILVUS_TOKEN": "your-zilliz-cloud-api-key" } } } } ``` </details> <details> <summary><strong>VS Code</strong></summary> The Claude Context MCP server can be used with VS Code through MCP-compatible extensions. Add the following configuration to your VS Code MCP settings: ```json { "mcpServers": { "claude-context": { "command": "npx", "args": ["-y", "@zilliz/claude-context-mcp@latest"], "env": { "OPENAI_API_KEY": "your-openai-api-key", "MILVUS_TOKEN": "your-zilliz-cloud-api-key" } } } } ``` </details> <details> <summary><strong>Cherry Studio</strong></summary> Cherry Studio allows for visual MCP server configuration through its settings interface. While it doesn't directly support manual JSON configuration, you can add a new server via the GUI: 1. Navigate to **Settings → MCP Servers → Add Server**. 2. Fill in the server details: - **Name**: `claude-context` - **Type**: `STDIO` - **Command**: `npx` - **Arguments**: `["@zilliz/claude-context-mcp@latest"]` - **Environment Variables**: - `OPENAI_API_KEY`: `your-openai-api-key` - `MILVUS_TOKEN`: `your-zilliz-cloud-api-key` 3. Save the configuration to activate the server. </details> <details> <summary><strong>Cline</strong></summary> Cline uses a JSON configuration file to manage MCP servers. To integrate the provided MCP server configuration: 1. Open Cline and click on the **MCP Servers** icon in the top navigation bar. 2. Select the **Installed** tab, then click **Advanced MCP Settings**. 3. In the `cline_mcp_settings.json` file, add the following configuration: ```json { "mcpServers": { "claude-context": { "command": "npx", "args": ["@zilliz/claude-context-mcp@latest"], "env": { "OPENAI_API_KEY": "your-openai-api-key", "MILVUS_TOKEN": "your-zilliz-cloud-api-key" } } } } ``` 4. Save the file. </details> <details> <summary><strong>Augment</strong></summary> To configure Claude Context MCP in Augment Code, you can use either the graphical interface or manual configuration. #### **A. Using the Augment Code UI** 1. Click the hamburger menu. 2. Select **Settings**. 3. Navigate to the **Tools** section. 4. Click the **+ Add MCP** button. 5. Enter the following command: ``` npx @zilliz/claude-context-mcp@latest ``` 6. Name the MCP: **Claude Context**. 7. Click the **Add** button. ------ #### **B. Manual Configuration** 1. Press Cmd/Ctrl Shift P or go to the hamburger menu in the Augment panel 2. Select Edit Settings 3. Under Advanced, click Edit in settings.json 4. Add the server configuration to the `mcpServers` array in the `augment.advanced` object ```json "augment.advanced": { "mcpServers": [ { "name": "claude-context", "command": "npx", "args": ["-y", "@zilliz/claude-context-mcp@latest"] } ] } ``` </details> <details> <summary><strong>Gemini CLI</strong></summary> Gemini CLI requires manual configuration through a JSON file: 1. Create or edit the `~/.gemini/settings.json` file. 2. Add the following configuration: ```json { "mcpServers": { "claude-context": { "command": "npx", "args": ["@zilliz/claude-context-mcp@latest"], "env": { "OPENAI_API_KEY": "your-openai-api-key", "MILVUS_TOKEN": "your-zilliz-cloud-api-key" } } } } ``` 3. Save the file and restart Gemini CLI to apply the changes. </details> <details> <summary><strong>Roo Code</strong></summary> Roo Code utilizes a JSON configuration file for MCP servers: 1. Open Roo Code and navigate to **Settings → MCP Servers → Edit Global Config**. 2. In the `mcp_settings.json` file, add the following configuration: ```json { "mcpServers": { "claude-context": { "command": "npx", "args": ["@zilliz/claude-context-mcp@latest"], "env": { "OPENAI_API_KEY": "your-openai-api-key", "MILVUS_TOKEN": "your-zilliz-cloud-api-key" } } } } ``` 3. Save the file to activate the server. </details> <details> <summary><strong>Other MCP Clients</strong></summary> The server uses stdio transport and follows the standard MCP protocol. It can be integrated with any MCP-compatible client by running: ```bash npx @zilliz/claude-context-mcp@latest ``` </details> ## Features - 🔌 **MCP Protocol Compliance**: Full compatibility with MCP-enabled AI assistants and agents - 🔍 **Hybrid Code Search**: Natural language queries using advanced hybrid search (BM25 + dense vector) to find relevant code snippets - 📁 **Codebase Indexing**: Index entire codebases for fast hybrid search across millions of lines of code - 🔄 **Incremental Indexing**: Efficiently re-index only changed files using Merkle trees for auto-sync - 🧩 **Intelligent Code Chunking**: AST-based code analysis for syntax-aware chunking with automatic fallback - 🗄️ **Scalable**: Integrates with Zilliz Cloud for scalable vector search, no matter how large your codebase is - 🛠️ **Customizable**: Configure file extensions, ignore patterns, and embedding models - ⚡ **Real-time**: Interactive indexing and searching with progress feedback ## Available Tools ### 1. `index_codebase` Index a codebase directory for hybrid search (BM25 + dense vector). **Parameters:** - `path` (required): Absolute path to the codebase directory to index - `force` (optional): Force re-indexing even if already indexed (default: false) - `splitter` (optional): Code splitter to use - 'ast' for syntax-aware splitting with automatic fallback, 'langchain' for character-based splitting (default: "ast") - `customExtensions` (optional): Additional file extensions to include beyond defaults (e.g., ['.vue', '.svelte', '.astro']). Extensions should include the dot prefix or will be automatically added (default: []) - `ignorePatterns` (optional): Additional ignore patterns to exclude specific files/directories beyond defaults (e.g., ['static/**', '*.tmp', 'private/**']) (default: []) ### 2. `search_code` Search the indexed codebase using natural language queries with hybrid search (BM25 + dense vector). **Parameters:** - `path` (required): Absolute path to the codebase directory to search in - `query` (required): Natural language query to search for in the codebase - `limit` (optional): Maximum number of results to return (default: 10, max: 50) - `extensionFilter` (optional): List of file extensions to filter results (e.g., ['.ts', '.py']) (default: []) ### 3. `clear_index` Clear the search index for a specific codebase. **Parameters:** - `path` (required): Absolute path to the codebase directory to clear index for ### 4. `get_indexing_status` Get the current indexing status of a codebase. Shows progress percentage for actively indexing codebases and completion status for indexed codebases. **Parameters:** - `path` (required): Absolute path to the codebase directory to check status for ## Contributing This package is part of the Claude Context monorepo. Please see: - [Main Contributing Guide](../../CONTRIBUTING.md) - General contribution guidelines - [MCP Package Contributing](CONTRIBUTING.md) - Specific development guide for this package ## Related Projects - **[@zilliz/claude-context-core](../core)** - Core indexing engine used by this MCP server - **[VSCode Extension](../vscode-extension)** - Alternative VSCode integration - [Model Context Protocol](https://modelcontextprotocol.io/) - Official MCP documentation ## License MIT - See [LICENSE](../../LICENSE) for details

Reviews

No reviews yet. Be the first to review this package!

Compatible With

Claude CodeCursorWindsurfContinue.dev

Details

Version
0.1.3
License
MIT
Category
ai
MCP Version
1.0
Published
8/3/2025
Updated
10/24/2025