PaddleOCR-X-High-Performance-Inference-MCP-Server-Suite
Verified Safeby williamIIliu
Overview
This server provides an intelligent document processing toolkit based on PaddleOCR 3.3 and PP-StructureV3, offering comprehensive functionalities like OCR, layout detection, table processing, formula recognition, and seal detection for various document types including PDFs.
Installation
paddleocr_mcp --pipeline PP-StructureV3 --ppocr_source local --host 0.0.0.0 --port 8090 --http --pipeline_config PaddleOCR.yaml --device gpuEnvironment Variables
- PYTHONPATH
- TENSORRT_ROOT
- LD_LIBRARY_PATH
- C_INCLUDE_PATH
- CPLUS_INCLUDE_PATH
- PADDLEOCR_MCP_PIPELINE
- PADDLEOCR_MCP_PPOCR_SOURCE
Security Notes
The server deploys an HTTP endpoint on `0.0.0.0`, which makes it accessible from all network interfaces. While this is common for services, it requires external security measures (e.g., firewalls, API gateways, authentication) to prevent unauthorized access. The project relies on specific versions of PaddlePaddle and CUDA/TensorRT, and mentions a manual workaround for an `ImportError` (`fused_rms_norm_ext`), which could introduce stability issues if not meticulously managed, but does not appear to be a direct security vulnerability. No hardcoded secrets or malicious patterns are evident in the provided code snippets.
Similar Servers
kreuzberg
Extracts text, tables, images, and metadata from a wide range of document formats (PDF, Office, images, HTML, etc.), with support for multiple OCR backends and an extensible plugin system. Can be run as a Micro-Agent Communication Protocol (MCP) server.
kreuzberg
Extracts text, tables, images, and metadata from 56 file formats including PDF, Office documents, and images. Supports multiple OCR backends, extensible plugins, and is designed for data preprocessing in AI/ML workflows.
pdf-reader-mcp
Provides a robust server for AI agents to extract text, images, and metadata from PDF documents, preserving content order for better comprehension.
academia_mcp
An MCP server providing tools for searching, fetching, analyzing, and reporting on scientific papers and datasets, often powered by LLMs.