Enables Claude and other LLMs to search through any published MkDocs documentation site using the Lunr.js search engine, allowing the AI to find and summarize relevant documentation for users.
A Model Context Protocol (MCP) server that provides search functionality for any MkDocs powered site. This server relies on the existing MkDocs search implementation using the Lunr.Js search engine.
Follow the installation instructions please follow the Model Context Protocol Quickstart For Claude Desktop users. You will need to add a section tothe MCP configuration file as follows:
{ "mcpServers": { "my-docs": { "command": "npx", "args": [ "-y", "@serverless-dna/mkdocs-mcp", "https://your-doc-site", "Describe what you are enabling search for to help your AI Agent" ] } } }
This project implements an MCP server that enables Large Language Models (LLMs) to search through any published mkdocs documentation site. It uses lunr.js for efficient local search capabilities and provides results that can be summarized and presented to users.
# Install dependencies pnpm install # Build the project pnpm build
The server can be run as an MCP server that communicates over stdio:
npx -y @serverless-dna/mkdocs-mcp https://your-doc-site.com
The server provides a search_docs
tool with the following parameters:
search
: The search query stringversion
: Optional version string (defaults to 'latest')pnpm build
pnpm test
During development you can run the MCP Server with Claude Desktop using the following configuration.
The configuration below shows running in windows claude desktop while developing using the Windows Subsystem for Linux (WSL). Mac or Linux environments you can run in a similar way.
The output is a bundled file which enables Node installed in windows to run the MCP server since all dependencies are bundled.
{ "mcpServers": { "powertools": { "command": "node", "args": [ "\\\\wsl$\\Ubuntu\\home\\walmsles\\dev\\serverless-dna\\mkdocs-mcp\\dist\\index.js", "Search online documentation" ] } } }
MIT
Discover shared experiences
Shared threads will appear here, showcasing real-world applications and insights from the community. Check back soon for updates!