This project includes a Model Context Protocol (MCP) server that exposes the indexed data through a standardized set of tools. This allows AI coding agents to interact with the indexed codebase in a structured way.
You must index your code base with the Semantic Code Search Indexer found here: https://github.com/elastic/semantic-code-search-indexer
The easiest way to run the MCP server is with Docker. The server is available on Docker Hub as simianhacker/semantic-code-search-mcp-server
.
To ensure you have the latest version of the image, run the following command before running the server:
docker pull simianhacker/semantic-code-search-mcp-server
This mode is useful for running the server in a containerized environment where it needs to be accessible over the network.
docker run --rm -p 3000:3000 \
-e ELASTICSEARCH_ENDPOINT=<your_elasticsearch_endpoint> \
simianhacker/semantic-code-search-mcp-server
Replace <your_elasticsearch_endpoint>
with the actual endpoint of your Elasticsearch instance.
This mode is useful for running the server as a local process that an agent can communicate with over stdin
and stdout
.
With Elasticsearch Endpoint:
docker run -i --rm \
-e ELASTICSEARCH_ENDPOINT=<your_elasticsearch_endpoint> \
simianhacker/semantic-code-search-mcp-server \
node dist/src/mcp_server/bin.js stdio
With Elastic Cloud ID:
docker run -i --rm \
-e ELASTICSEARCH_CLOUD_ID=<your_cloud_id> \
-e ELASTICSEARCH_API_KEY=<your_api_key> \
simianhacker/semantic-code-search-mcp-server \
node dist/src/mcp_server/bin.js stdio
The -i
flag is important as it tells Docker to run the container in interactive mode, which is necessary for the server to receive input from stdin
.
You can connect a coding agent to the server in either HTTP or STDIO mode.
HTTP Mode:
For agents that connect over HTTP, like the Gemini CLI, you can add the following to your ~/.gemini/settings.json
file:
{
"mcpServers": {
"Semantic Code Search": {
"trust": true,
"httpUrl": "http://localhost:3000/mcp/",
}
}
}
STDIO Mode:
For agents that connect over STDIO, you need to configure them to run the Docker command directly. Here's an example for the Gemini CLI in your ~/.gemini/settings.json
file:
{
"mcpServers": {
"SemanticCodeSearch": {
"command": "docker",
"args": [
"run",
"--rm",
"-i",
"-e", "ELASTICSEARCH_CLOUD_ID=<your_cloud_id>",
"-e", "ELASTICSEARCH_API_KEY=<your_api_key>",
"-e", "ELASTICSEARCH_INDEX=<your_index>",
"simianhacker/semantic-code-search-mcp-server",
"node", "dist/src/mcp_server/bin.js", "stdio"
]
}
}
}
Remember to replace the placeholder values for your Cloud ID, API key, and index name.
- Node.js (v20 or later)
- npm
- An running Elasticsearch instance (v8.0 or later) with the ELSER model downloaded and deployed.
git clone <repository-url>
cd semantic-code-search-mcp-server
npm install
Copy the .env.example
file and update it with your Elasticsearch credentials.
cp .env.example .env
The multi-threaded worker requires the project to be compiled to JavaScript.
npm run build
The MCP server can be run in two modes:
1. Stdio Mode:
This is the default mode. The server communicates over stdin
and stdout
.
npm run mcp-server
2. HTTP Mode: This mode is useful for running the server in a containerized environment like Docker.
npm run mcp-server:http
The server will listen on port 3000 by default. You can change the port by setting the PORT
environment variable.
You can also run the MCP server directly from the git repository using npx
. This is a convenient way to run the server without having to clone the repository.
Stdio Mode:
ELASTICSEARCH_ENDPOINT=http://localhost:9200 npx github:elastic/semantic-code-search-mcp-server
HTTP Mode:
PORT=8080 ELASTICSEARCH_ENDPOINT=http://localhost:9200 npx github:elastic/semantic-code-search-mcp-server http
Prompt | Description |
---|---|
StartInvestigation |
This prompt helps you start a "chain of investigation" to understand a codebase and accomplish a task. It follows a structured workflow that leverages the available tools to explore the code, analyze its components, and formulate a plan. |
Example:
/StartInvestigation --task="add a new route to the kibana server"
The MCP server provides the following tools:
Tool | Description |
---|---|
semantic_code_search |
Performs a semantic search on the code chunks in the index. This tool can combine a semantic query with a KQL filter to provide flexible and powerful search capabilities. |
map_symbols_by_query |
Query for a structured map of files containing specific symbols, grouped by file path. This is useful for finding all the symbols in a specific file or directory. Accepts an optional size parameter to control the number of files returned. |
symbol_analysis |
Analyzes a symbol and returns a report of its definitions, call sites, and references. This is useful for understanding the role of a symbol in the codebase. |
read_file_from_chunks |
Reads the content of a file from the index, providing a reconstructed view based on the most important indexed chunks. |
document_symbols |
Analyzes a file to identify the key symbols that would most benefit from documentation. This is useful for automating the process of improving the semantic quality of a codebase. |
Note: All of the tools accept an optional index
parameter that allows you to override the ELASTICSEARCH_INDEX
for a single query.
Configuration is managed via environment variables in a .env
file.
Variable | Description | Default |
---|---|---|
ELASTICSEARCH_CLOUD_ID |
The Cloud ID for your Elastic Cloud instance. | |
ELASTICSEARCH_API_KEY |
An API key for Elasticsearch authentication. | |
ELASTICSEARCH_INDEX |
The name of the Elasticsearch index to use. | semantic-code-search |