Software Engineering

Chat With Your ML Pipelines: Introducing the ZenML MCP Server

Alex Strick van Linschoten
Mar 10, 2025
5 mins

In a week where the Machine Learning community has been buzzing with unprecedented enthusiasm around the Model Context Protocol (MCP), our announcement of the ZenML MCP Server couldn't be more timely. As social media feeds, YouTube and developer forums overflow with MCP implementations and discussions, we're excited to introduce our contribution that brings the power of conversational AI directly to your ML pipelines—allowing you to interact with your infrastructure through natural language rather than commands or dashboards.

What We Built

We've developed a complete MCP server that integrates with MCP clients (like Claude Desktop, Cursor, and WindSurf). This integration allows you to communicate with your ML pipelines through natural language interfaces, making your workflows more accessible and interactive.

Currently, we've focused on providing read-only capabilities to ensure safe interactions with your metadata and pipelines. The server gives you access to information about users, stacks, pipelines, pipeline runs, pipeline steps, services, stack components, flavors, pipeline run templates, schedules, artifacts, service connectors, and step code. As an exception to the read-only approach, you can trigger new pipeline runs through existing run templates.

This implementation enables a range of practical use cases:

  • Querying your ZenML server using natural language
  • Generating analytics and visualizations from your pipeline run data
  • Creating detailed reports about pipeline performance
  • Investigating failing pipelines without switching contexts

Why It's Useful

The ZenML MCP Server offers numerous advantages for developers and teams:

  1. Natural Language Interaction: Query your ZenML metadata, code and logs using conversational language instead of memorizing CLI commands or navigating dashboard interfaces.
  2. Contextual Development: Get insights about failing pipelines or performance metrics without switching away from your development environment.
  3. Accessible Analytics: Generate custom reports and visualizations about your pipelines directly through conversation.
  4. Streamlined Workflows: Trigger pipeline runs via natural language requests when you're ready to execute.

The ability to "chat with your pipelines" opens up new possibilities for how teams can interact with their machine learning and LLM-based work, making complex operations more accessible to team members regardless of their familiarity with the underlying systems.

Understanding the MCP Landscape

The Model Context Protocol (MCP) was released in late November 2024 and it took a while for the protocol to really take off. More recently Mahesh Murag (Anthropic) held a workshop at the AI Engineering conference in which he spoke about how Anthropic saw future updates and expansions to the MCP.

Short-Term Developments

The engineering team are apparently working to implement these pieces at the moment, so we should expect to see them released in the not-so-distant future:

  • Remote Server Hosting: Moving from standard I/O to Server-Sent Events (SSE)
  • OAuth Support: Authentication mechanisms to connect with various services
  • MCP Registry: A centralized discovery service for finding and verifying MCP servers

Medium-Term Directions

As he moved into discussing things beyond the first half of 2025, it seemed like there were many more questions than necessarily fixed plans, but some of the items that they were looking into and exploring how to support were:

  • Stateful vs. Stateless Connections: Support for more flexible connection patterns
  • Streaming Improvements: Enhanced data streaming between components
  • Namespacing Solutions: Addressing tool naming conflicts across multiple servers
  • Proactive Server Behavior: Patterns for servers to initiate interactions

Of course, over the longer-term you can imagine that they want MCP to be thestandard protocol for agents and agents interactions. With this foundational layer for the development of agents in place, it'd be something a website or service would advertise and make available to their users.

The enthusiasm around MCP stems from its potential to standardize how AI systems interact with tools and services, creating network effects as more developers adopt the protocol and build compatible systems.

Engineering Choices and Tradeoffs

While quite a few implementations of MCP servers already exist, there nevertheless are some parts of the protocol that caused us to pause and I'd like to mention some of those here.

Read-Only for Now

We've deliberately chosen to make our MCP server primarily read-only for important safety reasons. While MCP clients typically request permission before executing tools, we've seen that users may become desensitized to these prompts after multiple interactions, potentially leading to unintended consequences with destructive actions.

By restricting our server to read-only operations (with the exception of triggering pipeline runs through templates), we've prioritized safety while still providing substantial utility. We're open to considering expanded capabilities based on community feedback and evolving best practices for secure MCP implementations. The code is also released openly and as it is run on your local machine you can of course add whatever write-based methods you'd like, at your own risk.

Local Running

Currently, the ZenML MCP server is designed to run locally, meaning you'll need to use it with a local MCP client like Claude Desktop or Cursor. This approach aligns with the current state of the MCP ecosystem, which is still developing standards for remote authentication and authorization.

Local execution offers advantages in terms of security and simplicity, though it does require some technical configuration. As the MCP standard evolves to support secure remote execution through SSE and other mechanisms, we anticipate exploring options for hosted versions of the ZenML MCP server.

Installation and Dependencies

We've configured the server to work with uv for dependency management, which provides a reliable way to reproduce the necessary Python environment. While this represents an additional component to install, it significantly simplifies the overall setup process by handling environment configuration automatically.

The installation requires minimal technical effort: cloning the repository, configuring your connection details, and setting up the integration with your preferred MCP client. Detailed instructions are available in the GitHub repository.

Getting Started

The setup process for the ZenML MCP Server is straightforward:

  • Prerequisites:
    • Access to a ZenML Cloud server
    • UV installed locally
    • A local clone of the repository
  • Configuration:
    • Create an MCP config file with your ZenML server details
    • Configure your preferred MCP client (Claude Desktop or Cursor)
  • Usage:
    • Start interacting with your ZenML infrastructure through natural language

For detailed setup instructions, please refer to the GitHub repository.

Example Prompts to Try

Once you've set up the ZenML MCP Server, you're only limited by your imagination. To give you an idea of some of the kinds of prompts you might want to write, here are some to play around with. The first three have screenshots of the kind of output you might expect to see.

Pipeline Analysis Report

"Can you write me a report (as a markdown artifact) about the simple_pipeline and tell the story of the history of its runs, which were successful etc., and what stacks worked, which didn't, as well as some performance metrics + recommendations?"

Comparative Pipeline Analysis

"Could you analyze all our ZenML pipelines and create a comparison report (as a markdown artifact) that highlights differences in success rates, average run times, and resource usage? Please include a section on which stacks perform best for each pipeline type."

Stack Component Analysis

"Please generate a comprehensive report or dashboard on our ZenML stack components, showing which ones are most frequently used across our pipelines. Include information about version compatibility issues and performance variations."

Run Template Analysis

"Could you analyze our ZenML pipeline run templates and create a markdown report that shows how frequently each template is used, their average success rates, and execution times?"

Future Directions

We're excited about the potential evolution of the ZenML MCP Server:

  • Hosted Servers: Depending on how Anthropic and the MCP ecosystem develop, we may offer hosted versions to eliminate local setup requirements
  • Write Actions: Potential expansion to include safe write operations based on community feedback
  • Extended Capabilities: Further integration with evolving MCP features and protocols

The development roadmap will be heavily influenced by user feedback and the broader MCP ecosystem's evolution.

Get Involved

We invite you to try the ZenML MCP Server and share your experiences with us through Slack. We're particularly interested in:

  • Whether you need additional write actions (creating stacks, registering components, etc.)
  • Examples of how you're using the server in your workflows
  • Suggestions for additional features or improvements

Contributions and pull requests to the core repository are always welcome!

By bringing natural language interfaces to your ML pipelines, we're taking another step toward making machine learning operations more accessible, efficient, and integrated with your development workflow. We look forward to seeing how you incorporate this capability into your ML projects.

Looking to Get Ahead in MLOps & LLMOps?

Subscribe to the ZenML newsletter and receive regular product updates, tutorials, examples, and more articles like this one.
We care about your data in our privacy policy.