Resources
The MCP server exposes resources that agents can read on demand for detailed documentation.| Resource URI | Description |
|---|---|
datris://pipeline-config-reference | Complete reference for building pipeline configurations — all source types, data quality rules, transformations, and destination types with JSON examples |
Available Tools
Pipeline Management
| Tool | Description |
|---|---|
list_pipelines | List all registered pipeline configurations |
get_pipeline | Get a specific pipeline configuration by name |
create_pipeline | Create a pipeline from sample data (base64-encoded). Schema is auto-detected. Specify destination type. |
delete_pipeline | Delete a pipeline and its destination data |
upload_data | Upload data (base64-encoded) to a pipeline for processing (returns pipeline token) |
get_job_status | Get job status by pipeline token or pipeline name |
kill_job | Kill a running job by pipeline token |
profile_data | AI-profile data (base64-encoded) with summary stats and suggested DQ rules |
get_version | Get pipeline server version |
check_service_health | Check which backend services are up, down, or not configured (slow — use for diagnostics only) |
Vector Database Search
Semantic search across any of the pipeline’s supported vector databases. Each tool takes a natural language query and returns the most similar document chunks with scores and metadata.| Tool | Description |
|---|---|
search_qdrant | Search a Qdrant collection |
search_weaviate | Search a Weaviate class |
search_milvus | Search a Milvus collection |
search_chroma | Search a Chroma collection |
search_pgvector | Search a pgvector PostgreSQL table |
Database Queries
Read-only queries against the pipeline’s backend databases.| Tool | Description |
|---|---|
query_postgres | Execute a read-only SQL SELECT query against PostgreSQL |
query_mongodb | Query a MongoDB collection with filter and projection |
query_natural | Ask a question in natural language — AI generates and executes SQL |
Metadata Discovery
Explore the structure of PostgreSQL, MongoDB, and vector databases managed by the platform. Use these tools to understand what data is available before writing queries or running searches.| Tool | Description |
|---|---|
list_postgres_databases | List all PostgreSQL databases |
list_postgres_schemas | List schemas in a PostgreSQL database |
list_postgres_tables | List tables in a schema (supports vector-only filter) |
list_postgres_columns | List columns and types for a specific table |
list_mongodb_databases | List all MongoDB databases |
list_mongodb_collections | List collections (optionally filtered by database) |
list_qdrant_collections | List all collections in Qdrant |
list_weaviate_classes | List all classes in Weaviate |
list_milvus_collections | List all collections in Milvus |
list_chroma_collections | List all collections in Chroma |
list_pgvector_collections | List all pgvector tables in PostgreSQL |
AI
| Tool | Description |
|---|---|
ai_answer | Answer a question using AI based on provided context (RAG) |
Configuration
| Tool | Description |
|---|---|
upload_config | Upload a JSON Schema config file (base64-encoded content) |
update_secret | Update an AI provider secret (anthropic, openai, ollama, embedding) to configure API keys |
Setup
Docker (automatic)
The MCP server starts automatically withdocker-compose up in SSE mode on port 3000. No additional setup required.
Local (for Claude Desktop / Claude Code)
The MCP server is published on PyPI. Useuvx to run it directly:
Transport Modes
| Mode | Use Case | Command |
|---|---|---|
| stdio | Claude Desktop, Claude Code, local agents | python server.py |
| SSE | Docker, remote agents, web clients | python server.py --sse --port 3000 |
Configuring Claude Desktop
Add to~/Library/Application Support/Claude/claude_desktop_config.json:
Configuring Claude Code
Add to.mcp.json in your project root:
Environment Variables
| Variable | Default | Description |
|---|---|---|
PIPELINE_URL | http://localhost:8080 | Pipeline server URL |
PIPELINE_API_KEY | (empty) | API key if pipeline has key validation enabled |
Example Agent Workflows
Profile and ingest a CSV file
An AI agent could autonomously:- Profile the data —
profile_datawith the CSV file - Review suggested rules — agent reads the AI-suggested DQ rules
- Create the pipeline —
create_pipelinewith the profiled config - Upload the file —
upload_datato trigger processing - Monitor status —
get_job_statusto track completion
Build and query a RAG knowledge base
- Create pipeline —
create_pipelinewith Qdrant/Weaviate/Milvus/pgvector destination - Upload documents —
upload_datafor each PDF/document - Monitor —
get_job_statusuntil all documents are processed - Search —
search_qdrantto find relevant chunks - Answer —
ai_answerwith the retrieved chunks as context and the user’s question
Discover and query data
- List databases —
list_postgres_databasesto see available databases - List schemas —
list_postgres_schemasto explore a database - List tables —
list_postgres_tablesto find relevant tables - Inspect columns —
list_postgres_columnsto understand table structure - Query —
query_postgreswith a well-formed SELECT query
Cross-database analysis
- Search documents —
search_pgvectorfor relevant financial document chunks - Query structured data —
query_postgresto get related financial metrics - Combine — agent merges unstructured + structured data in its response
AI validation and transformation
- Create pipeline with AI rules —
create_pipelinewithcodegen_rulefor validation and/orcodegen_transformfor transformation - Upload data —
upload_datato process data through the AI-powered pipeline - Datris generates Python scripts from your instructions and runs them in the container
Automated data quality monitoring
- List pipelines —
list_pipelinesto discover all registered pipelines - Upload new data —
upload_datawith latest data files - Check results —
get_job_statusto see DQ failures - Diagnose — AI reads error explanations and suggests fixes
