MCP-rag-with-Chromadb
by rexnzm
Overview
Retrieval-Augmented Generation (RAG) server for multi-format document ingestion and vector database integration, supporting Ollama or OpenAI embeddings.
Installation
python server.pyEnvironment Variables
- EMBEDDING_PROVIDER
- OPENAI_API_KEY
- OPENAI_EMBED_MODEL
- OLLAMA_EMBED_MODEL
- OLLAMA_BASE_URL
Security Notes
The primary security concern arises from the distribution method outlined in the README, which instructs users to download and run executables (.exe, .dmg) or archives (tar.gz) directly from raw GitHub URLs. This bypasses secure software distribution practices and poses a significant supply chain risk, making it unsafe to run by following the project's own setup instructions. Additionally, the 'ingest_document' tool can download files from arbitrary URLs, which could be exploited for denial-of-service or malicious content download if the server were exposed to untrusted inputs. However, no 'eval' or explicit hardcoded secrets are found in the provided Python source code, and network calls are to standard endpoints for Ollama or OpenAI.
Similar Servers
haiku.rag
Opinionated agentic RAG powered by LanceDB, Pydantic AI, and Docling to provide hybrid search, intelligent QA, and multi-agent research over user-provided documents, accessible via CLI, Python API, Web App, TUI, or as an MCP server for AI assistants.
flexible-graphrag
The Flexible GraphRAG MCP Server integrates document processing, knowledge graph building, hybrid search, and AI query capabilities via the Model Context Protocol (MCP) for clients like Claude Desktop and MCP Inspector.
rag-server-mcp
Provides Retrieval Augmented Generation (RAG) capabilities to Model Context Protocol (MCP) clients by indexing project documents and retrieving relevant content for LLMs.
RAG-CHROMA-MCP-SERVER
Serves a Retrieval-Augmented Generation (RAG) system via HTTP/SSE using the Model Context Protocol (MCP), primarily for integration with AI assistants like Claude Desktop.