Related MCP Server Resources

Explore more AI models, providers, and integration options:

  • Explore AI Models
  • Explore AI Providers
  • Explore MCP Servers
  • LangDB Pricing
  • Documentation
  • AI Industry Blog
  • Figma MCP Server
  • AgentCraft MCP Server
  • Git Spice Help MCP Server
  • Notion Knowledge Base MCP Server
  • DeepSource MCP Server
Back to MCP Servers
RAG MCP server

RAG MCP server

Public
proofofsid/rag-mcp

Implements a RAG workflow that integrates with any custom knowledge base and can be triggered directly from the Cursor IDE.

Verified
python
0 tools
May 30, 2025
Updated May 30, 2025

RAG-MCP Server

A general-purpose Retrieval-Augmented Generation (RAG) server using the Model Control Protocol (MCP), designed to be tested with RISC Zero's Bonsai documentation.

Overview

This project implements a RAG server that:

  • Uses MCP (Model Control Protocol) for standardized communication
  • Implements RAG (Retrieval-Augmented Generation) workflow for document querying
  • Can be tested with RISC Zero's Bonsai documentation
  • Supports local LLM integration through Ollama

Features

  • Document ingestion and indexing
  • Semantic search capabilities
  • Local LLM integration
  • MCP protocol compliance
  • RISC Zero Bonsai documentation support

Prerequisites

  • Python 3.12+
  • Ollama (for local LLM support)
  • Poetry (for dependency management)

Installation

  1. Install Python dependencies:
poetry install
  1. Install and start Ollama:
# Install Ollama brew install ollama # for macOS # or curl -fsSL https://ollama.com/install.sh | sh # for Linux # Start Ollama service ollama serve
  1. Pull the required model:
ollama pull llama2

Usage

  1. Start the MCP server:
poetry run python mcp_server.py
  1. The server will:

    • Initialize the LLM and embedding model
    • Ingest documents from the data directory
    • Process queries using the RAG workflow
  2. Test with RISC Zero Bonsai docs:

    • Place RISC Zero Bonsai documentation in the data/ directory
    • Query the server about Bonsai features and implementation

Project Structure

  • mcp_server.py: Main server implementation
  • rag.py: RAG workflow implementation
  • data/: Directory for document ingestion
  • storage/: Vector store and document storage
  • start_ollama.sh: Script to start Ollama service

Testing with RISC Zero Bonsai

The server is configured to work with RISC Zero's Bonsai documentation. You can:

  1. Add Bonsai documentation to the data/ directory
  2. Query about Bonsai features, implementation details, and usage
  3. Test the RAG workflow with Bonsai-specific questions

Made with ❤️ by proofofsid

Publicly Shared Threads0

Discover shared experiences

Shared threads will appear here, showcasing real-world applications and insights from the community. Check back soon for updates!

Share your threads to help others
Related MCPs5
  • Figma MCP Server
    Figma MCP Server

    Gives AI-powered coding tools like Cursor, Windsurf, and Cline access to Figma design files, enablin...

    Added May 30, 2025
  • AgentCraft MCP Server
    AgentCraft MCP Server

    Integrates with the AgentCraft framework to enable secure communication and data exchange between AI...

    Added May 30, 2025
  • Git Spice Help MCP Server
    Git Spice Help MCP Server

    A Model Context Protocol server that integrates with Cursor IDE to provide real-time git-spice docum...

    1 tools
    Added May 30, 2025
  • Notion Knowledge Base MCP Server
    Notion Knowledge Base MCP Server

    An MCP server that connects to your Notion knowledge base, allowing you to query and retrieve inform...

    Added May 30, 2025
  • DeepSource MCP Server
    DeepSource MCP Server

    A Model Context Protocol server that integrates with DeepSource to provide AI assistants with access...

    9 tools
    Added May 30, 2025