dracor-mcp-evaluation
Verified Safeby dracor-org
Overview
Automated evaluation of LLM tool-calling capabilities against the DraCor API for digital humanities research.
Installation
python automatic_prompting.pyEnvironment Variables
- ANTHROPIC_API_KEY
Security Notes
The system interacts with external APIs (Anthropic, DraCor). API keys (Anthropic) are expected to be managed via environment variables, which is a good security practice. The use of a JavaScript REPL is noted, but this is part of the simulated environment being evaluated by the `automatic_prompting.py` script, not an inherent vulnerability in the evaluation script itself. No direct 'eval' or obvious malicious patterns are present in the provided source code snippets for the evaluation script.
Similar Servers
mcp-interviewer
A Python CLI tool to evaluate Model Context Protocol (MCP) servers for agentic use-cases, by inspecting capabilities, running functional tests, and providing LLM-as-a-judge evaluations.
mcpbr
A benchmark runner for evaluating Model Context Protocol (MCP) servers by comparing LLM agent performance with and without MCP tools on software engineering tasks.
mcp-jest
A testing framework for Model Context Protocol (MCP) servers, allowing automated validation of AI agent tools, resources, and prompts.
mcp-k6
The mcp-k6 server provides an experimental Model Context Protocol (MCP) interface for k6, enabling script validation, test execution, documentation browsing, and guided script generation for performance testing workflows.