Altana, a global supply chain intelligence company, faced challenges in efficiently deploying and managing multiple GenAI models for diverse customer use cases. By implementing Databricks Mosaic AI platform, they transformed their ML lifecycle management, combining custom deep learning models with fine-tuned LLMs and RAG workflows. This led to 20x faster model deployment times and 20-50% performance improvements, while maintaining data privacy and governance requirements across their global operations.
Altana represents an interesting case study in implementing a sophisticated LLMOps infrastructure for supply chain intelligence applications. The company operates globally with nearly 200 employees serving over a thousand customers, focusing on making global commerce more transparent and reliable through AI-powered insights.
Their LLMOps journey showcases a sophisticated approach to combining multiple AI technologies in production, which they term a "compound AI systems" approach. This strategy involves several key components:
**Core Technical Architecture:**
* Integration of custom deep learning models for initial processing
* Implementation of fine-tuned agent workflows with RAG (Retrieval Augmented Generation)
* Use of RLHF (Reinforcement Learning from Human Feedback) for continuous model improvement
* Utilization of multiple open-source LLMs including DBRX, Llama 3, and Phi-3
* Integration with proprietary datasets for enhanced RAG workflows
**Production Use Cases:**
The system handles several sophisticated production applications:
* Automated generation of tax and tariff classifications for cross-border trading
* Production of detailed legal justification write-ups
* Enhancement of their global supply chain knowledge graph
* Privacy-preserving integration of customer data
* Real-time decision support for supply chain dependencies
**LLMOps Infrastructure and Tools:**
The implementation leverages several key components through Databricks Mosaic AI:
* Managed MLflow for comprehensive ML lifecycle management
* Model Serving capabilities for deployment
* Delta Lake for real-time data ingestion and management
* Unity Catalog for data governance and privacy controls
**Data Management and Privacy:**
A particularly noteworthy aspect of their LLMOps implementation is the strong focus on data governance and privacy. They maintain a federated architecture where all customer data remains separate and private, using Unity Catalog to enforce these governance requirements while maintaining operational efficiency.
**Monitoring and Optimization:**
Their LLMOps system includes sophisticated tracking of user interactions:
* Comprehensive monitoring of input/output decisions
* Performance tracking across different workflow steps
* Cost optimization analysis
* Latency monitoring
* Model performance evaluation
**Deployment and Scaling Strategy:**
The company adopted a flexible deployment approach that allows them to:
* Deploy across multiple cloud environments
* Maintain control over model swapping and regional deployments
* Optimize specific components as needed
* Scale while managing total cost of ownership
**Integration and Collaboration:**
A key success factor in their LLMOps implementation was the ability to enable collaboration across different teams:
* Data ingestion teams
* Data engineering teams
* ML teams
* Software development teams
The platform enables these teams to work together seamlessly in the same environment, significantly reducing coordination overhead and accelerating development cycles.
**Results and Metrics:**
The implementation has yielded significant measurable improvements:
* 20x faster deployment of GenAI models into production
* 20-50% improvement in model performance across workloads
* Expanded model catalog
* Improved time to value for customers
**Challenges and Solutions:**
Before implementing their current LLMOps infrastructure, Altana faced several challenges:
* Resource drain from building boilerplate infrastructure
* Difficulty in rapid iteration of approaches
* Complexity in fine-tuning data and models for specific customers
* Challenges in measuring R&D and live performance across cloud deployments
Their solution addressed these challenges through:
* Unified platform for data, models, hosting, and performance management
* Integrated tooling for the entire ML lifecycle
* Automated evaluation tools
* Streamlined deployment processes
**Future Directions:**
The company is continuing to expand their LLMOps implementation:
* Deeper integration of agentic workflows in ETL processes
* Enhanced enrichment, validation, and normalization capabilities
* Expanded use of AI across their product surface area
**Critical Analysis:**
While the case study demonstrates impressive results, it's worth noting that the implementation heavily relies on Databricks' ecosystem. This could potentially create some vendor lock-in, though the use of open-source models and tools helps mitigate this risk. The company appears to have made a conscious trade-off between the benefits of a unified platform and maintaining flexibility through open-source components.
The approach to privacy and governance is particularly noteworthy, as it demonstrates how to implement LLMOps at scale while maintaining strict data separation and security requirements. This is increasingly important as organizations deploy AI systems that handle sensitive business data.
Their "compound AI systems" approach, combining multiple model types and learning approaches, represents a sophisticated implementation of LLMOps that goes beyond simple model deployment. This demonstrates the maturity of their AI operations and provides a valuable template for organizations looking to implement complex AI systems in production.
Start your new ML Project today with ZenML Pro
Join 1,000s of members already deploying models with ZenML.