pinocchio
Verified Safeby tomblancdev
Overview
An MCP server that spawns autonomous Claude Code agents in isolated Docker containers for secure code analysis, development, and modification.
Installation
exec docker compose run --rm -T --service-ports mcp-serverEnvironment Variables
- PROJECTS_ROOT
- HOST_CLAUDE_DIR
- HOST_CONFIG_DIR
- HOST_GH_CONFIG
- ABSOLUTE_MAX_TIMEOUT
- MAX_CONCURRENT_AGENTS
- MAX_SPAWNS_PER_MINUTE
- DOCKER_HOST
Security Notes
The project demonstrates a robust defense-in-depth security model. Key mitigations include a Docker Socket Proxy to block dangerous operations (BUILD, COMMIT, EXEC, VOLUMES, SECRETS), a workspace allowlist with symlink resolution (fs.realpath) and explicit path/glob validation (including a dangerous glob blocklist), read-only workspace mounts by default with granular write access, container hardening (non-root, CAP_DROP ALL, memory/CPU limits, no-new-privileges), secure handling of GitHub tokens via temporary files with restrictive permissions, rate limiting for concurrent agents and spawn frequency, and comprehensive input validation for container names, tasks, and agent IDs. Audit logging tracks security-relevant events. There are no obvious 'eval' or obfuscation patterns. The main remaining considerations are the inherent prompt injection risk common to LLMs operating in 'YOLO' mode and that agent containers run with `ReadonlyRootfs: false` (necessary for dynamic package installation) and default network access (needed for LLM API communication).
Similar Servers
claude-code-mcp
Acts as an MCP server to enable LLMs to run Claude Code CLI in one-shot mode, bypassing permissions for complex coding, file system, Git, and terminal operations.
claude-codex-settings
A comprehensive toolkit and configuration for developing Claude Code plugins, integrating various external services and APIs, and enhancing AI-assisted coding workflows.
consult-llm-mcp
An MCP server that allows AI agents like Claude Code to consult stronger, more capable AI models (e.g., GPT-5.2, Gemini 3.0 Pro) for complex code analysis, debugging, and architectural advice.
multi_mcp
A multi-model AI orchestration server providing automated code review, security analysis, and general LLM-powered assistance through the Model Context Protocol (MCP).