Integrations
Docker
and
ZenML logo in purple, representing machine learning pipelines and MLOps framework.
Effortlessly Run ZenML Pipelines in Isolated Docker Containers
The image is blank. No elements are visible for description or keyword inclusion.
Docker
All integrations

Docker

Effortlessly Run ZenML Pipelines in Isolated Docker Containers
Add to ZenML
Category
Orchestrator
COMPARE
related resources
No items found.

Effortlessly Run ZenML Pipelines in Isolated Docker Containers

Integrate ZenML with Docker to execute your ML pipelines in isolated environments locally. This integration simplifies debugging and ensures consistent execution across different systems.

Features with ZenML

  • Isolated Pipeline Execution: Run each step of your ZenML pipeline in a separate Docker container, ensuring isolation and reproducibility.
  • Local Debugging: Debug issues that occur when running pipelines in Docker containers without the need for remote infrastructure.
  • Consistent Environments: Maintain consistent execution environments across different systems by leveraging Docker containers.
  • Easy Setup: Seamlessly integrate Docker with ZenML using the built-in local Docker orchestrator.

Main Features

  • Containerization of applications
  • Isolation of processes and dependencies
  • Portability across different systems
  • Efficient resource utilization
  • Reproducibility of environments

How to use ZenML with
Docker

from zenml import step, pipeline
from zenml.orchestrators.local_docker.local_docker_orchestrator import (
    LocalDockerOrchestratorSettings,
)

@step
def preprocess_data():
    # Preprocessing logic here
    pass

@step
def train_model():
    # Model training logic here
    pass

settings = {
    "orchestrator.local_docker": LocalDockerOrchestratorSettings(
        run_args={"cpu_count": 2}
    )
}

@pipeline(settings=settings)
def ml_pipeline():
    data = preprocess_data()
    train_model(data)

if __name__ == "__main__":
    ml_pipeline()

This code example demonstrates how to create a simple ZenML pipeline that runs each step in a Docker container using the local Docker orchestrator. The pipeline consists of two steps: preprocess_data and train_model. The LocalDockerOrchestratorSettings is used to specify additional configuration, such as the CPU count for the Docker containers. Finally, the pipeline is executed by calling ml_pipeline().

Additional Resources
Read the full documentation on the Local Docker Orchestrator
Learn how to enable GPU support with the Local Docker Orchestrator
Explore the ZenML SDK documentation for LocalDockerOrchestrator

Effortlessly Run ZenML Pipelines in Isolated Docker Containers

Integrate ZenML with Docker to execute your ML pipelines in isolated environments locally. This integration simplifies debugging and ensures consistent execution across different systems.
Docker

Start Your Free Trial Now

No new paradigms - Bring your own tools and infrastructure
No data leaves your servers, we only track metadata
Free trial included - no strings attached, cancel anytime
Dashboard displaying machine learning models, including versions, authors, and tags. Relevant to model monitoring and ML pipelines.

Connect Your ML Pipelines to a World of Tools

Expand your ML pipelines with Apache Airflow and other 50+ ZenML Integrations
Seldon
Comet
Apache Airflow
Neptune
Tekton
Azure Blob Storage
Google Cloud Storage (GCS)
Kaniko
TensorBoard
Argilla
Google Cloud Vertex AI Pipelines