thinkingcap
Verified Safeby Infatoshi
Overview
A multi-agent research MCP server that runs multiple LLM providers in parallel and synthesizes their responses to a given query.
Installation
npx -y thinkingcap openrouter:moonshotai/kimi-k2-thinking groq:moonshotai/kimi-k2-instruct-0905 cerebras:zai-glm-4.6 xai:grok-4-fastEnvironment Variables
- OPENAI_API_KEY
- OPENROUTER_API_KEY
- GROQ_API_KEY
- CEREBRAS_API_KEY
- XAI_API_KEY
- ANTHROPIC_API_KEY
- GOOGLE_API_KEY
Security Notes
API keys are correctly managed via environment variables and are not hardcoded. The system uses LLMs to generate structured data (questions) that are then JSON parsed; while this introduces a potential risk if an LLM deviates maliciously from the expected format, the prompts explicitly guide the LLM to return only a JSON array of strings, mitigating the risk. Web search is performed against DuckDuckGo, a legitimate search engine, via direct HTTP requests. There are no detected uses of `eval` or similar dangerous functions on untrusted inputs.
Similar Servers
gpt-researcher
The GPT Researcher MCP Server enables AI assistants to conduct comprehensive web research and generate detailed, factual, and unbiased reports. It supports multi-agent workflows, local document analysis, and integration with external tools via the Machine Conversation Protocol (MCP) for various research tasks.
deep-research
Generate comprehensive, AI-powered deep research reports, leveraging various LLMs and web search engines, with local knowledge base integration and report artifact editing.
mcp-omnisearch
Provides a unified interface for various search, AI response, content processing, and enhancement tools via Model Context Protocol (MCP).
Google-Search-MCP-Server
This MCP server enhances Google search with AI-powered research synthesis, content extraction, source quality assessment, and deduplication, designed to be used by large language models (LLMs) like Claude.