tiny_chat
by to-aoki
Overview
A RAG-enabled chat application that integrates with various LLM backends (OpenAI, Ollama, vLLM) and a Qdrant vector database, offering web search capabilities and an OpenAI-compatible API.
Installation
tiny-chat-mcpEnvironment Variables
- DB_CONFIG
- LLM_CONFIG
Security Notes
The application includes a `webbrowser.open` call in the Streamlit UI which can open arbitrary local files or URLs. If `source_name` (from uploaded files, web search, or database metadata) contains a malicious path (e.g., `file://`), it could lead to local machine compromise. FastAPI endpoints (e.g., `tiny_chat.api.compat_openai`, `tiny_chat.api.rest`, `tiny_chat.api.continue_context_provider`, `tiny_chat.mcp.search_mcp`) are configured to listen on `0.0.0.0` by default, exposing them to all network interfaces. This requires careful network configuration (e.g., firewall, reverse proxy) to prevent unauthorized access. File system operations for directory processing in `tiny_chat.database.components.registration.py` could also pose a risk if the input path is not strictly controlled in an exposed environment.
Similar Servers
flexible-graphrag
The Flexible GraphRAG MCP Server integrates document processing, knowledge graph building, hybrid search, and AI query capabilities via the Model Context Protocol (MCP) for clients like Claude Desktop and MCP Inspector.
qdrant-loader
A Model Context Protocol (MCP) server that provides advanced Retrieval-Augmented Generation (RAG) capabilities to AI development tools by bridging a QDrant knowledge base for intelligent, context-aware search.
Docker_MCPGUIApp
This repository provides a starter template for building full-stack AI assistants that integrate with real-world tools using Docker MCP Gateway and a Large Language Model.
RagThisCode
Set up a RAG (Retrieval-Augmented Generation) system to chat with the code of any public or private GitHub repository.