ZenML
Blog AI Engineering vs ML Engineering: Evolving Roles in the GenAI Era
MLOps 2 mins

AI Engineering vs ML Engineering: Evolving Roles in the GenAI Era

The rise of Generative AI has shifted the roles of AI Engineering and ML Engineering, with AI Engineers integrating generative AI into software products. This shift requires clear ownership boundaries and specialized expertise. A proposed solution is layer separation, separating concerns into two distinct layers: Application (AI Engineers/Software Engineers), Frontend development, Backend APIs, Business logic, User experience, and ML (ML Engineers). This allows AI Engineers to focus on user experience while ML Engineers optimize AI systems.

AI Engineering vs ML Engineering: Evolving Roles in the GenAI Era
On this page

The rise of Generative AI has introduced new complexities in how we structure engineering teams and ownership of AI products. Having worked on ZenML, I’ve observed the transition from a clear MLOps/ML Engineering divide to a more nuanced landscape that now includes AI Engineers.

The Current State

Traditional ML Engineering focused on building and deploying classical machine learning models, with clear ownership boundaries between ML Engineers, MLOps Engineers, and Data Scientists. AI Engineers have emerged as a new role, primarily focused on integrating generative AI into software products through API calls to models like GPT-4 or Claude.

While AI Engineers often come from full-stack backgrounds with strong JavaScript/TypeScript and Python skills, this creates challenges at scale. What starts as simple API calls to OpenAI quickly evolves into complex data and ML problems that traditional software engineers aren’t equipped to handle.

The Ownership Problem

Take RAG applications as an example. They require:

  • Document chunking strategies
  • Data ingestion pipelines
  • Refresh mechanisms
  • Experiment tracking (e.g. MLflow)
  • LLM observability (Langsmith /)

These aspects align more closely with ML engineering skills than traditional software engineering. While software engineers can learn these skills, enterprise environments need clear ownership boundaries and specialized expertise.

A Proposed Solution: Layer Separation

__wf_reserved_inherit

The most effective approach is separating concerns into two distinct layers:

1. Application Layer (AI Engineers/Software Engineers)

  • Frontend development (React, Vue, etc.)
  • Backend APIs (FastAPI, Express)
  • Business logic
  • User experience

2. ML Layer (ML Engineers)

  • Prompt management / fine-tuning / embeddings strategy
  • Data pipeline management
  • Evaluation frameworks
  • MLOps infrastructure

These layers communicate through well-defined interfaces like REST APIs or pub/sub patterns, but maintain independent lifecycles and ownership.

__wf_reserved_inherit

Example: Enterprise RAG Platform

Let’s examine how this layer separation works in practice. Consider building an internal platform where employees can create their own RAG-powered chatbots without writing code.

Consider building an internal no-code platform for employees to create RAG-powered chatbots:

__wf_reserved_inherit

The Application team owns the user experience:

  • Building the React dashboard where users drag and drop documents
  • Implementing the FastAPI backend for business logic and user management
  • Creating intuitive interfaces for model and agent configuration

Meanwhile, the ML team handles the AI infrastructure:

  • Running robust ingestion pipelines for different document types
  • Fine-tuning models for specific use cases
  • Building evaluation frameworks to ensure quality
  • Managing the agent orchestration layer

An internal platform team (MLOps / AI Platform) provides the bridge between these layers, maintaining shared infrastructure and ensuring smooth integration. This setup lets the AI Engineers focus on user experience while ML Engineers optimize the underlying AI systems.

Looking Forward

While “everyone will be an AI Engineer” is a common refrain, specialized roles remain crucial for building robust AI systems at scale. The key is establishing clear boundaries between application development and ML infrastructure, allowing each team to focus on their core competencies while working toward a common goal.

The industry needs to move away from bespoke solutions and toward standardized patterns for building AI-powered applications. This starts with recognizing the distinct skill sets required and structuring teams accordingly.

Start deploying AI workflows in production today

Enterprise-grade AI platform trusted by thousands of companies in production

Continue Reading

Chat With Your ML Pipelines: Introducing the ZenML MCP Server

Chat With Your ML Pipelines: Introducing the ZenML MCP Server

Discover the new ZenML MCP Server that brings conversational AI to ML pipelines. Learn how this implementation of the Model Context Protocol allows natural language interaction with your infrastructure, enabling query capabilities, pipeline analytics, and run management through simple conversation. Explore current features, engineering decisions, and future roadmap for this timely addition to the rapidly evolving MCP ecosystem.

Everything you ever wanted to know about LLMOps Maturity Models

Everything you ever wanted to know about LLMOps Maturity Models

As organizations rush to adopt generative AI, several major tech companies have proposed maturity models to guide this journey. While these frameworks offer useful vocabulary for discussing organizational progress, they should be viewed as descriptive rather than prescriptive guides. Rather than rigidly following these models, organizations are better served by focusing on solving real problems while maintaining strong engineering practices, building on proven DevOps and MLOps principles while adapting to the unique challenges of GenAI implementation.