Addverb developed an AI-powered voice control system for AGV (Automated Guided Vehicle) maintenance that enables warehouse workers to communicate with robots in their native language. The system uses a combination of edge-deployed Llama 3 and cloud-based ChatGPT to translate natural language commands from 98 different languages into AGV instructions, significantly reducing maintenance downtime and improving operational efficiency.
Addverb, a robotics company founded in 2016, has implemented an innovative LLMOps solution to address the growing challenge of AGV fleet maintenance in warehouse environments. This case study demonstrates a practical application of large language models in industrial robotics, combining edge computing and cloud services to create a robust production system.
### System Overview and Business Context
The company faced a significant challenge in warehouse operations where maintenance issues with Automated Guided Vehicles (AGVs) required specialized engineers, leading to potential delays and operational inefficiencies. To address this, they developed an AI-powered voice control system that allows any warehouse worker, regardless of technical expertise or language background, to communicate with and manage AGV maintenance tasks.
### Technical Architecture and LLM Implementation
The solution employs a hybrid approach to LLM deployment, utilizing both edge and cloud computing:
* Edge Deployment:
* Utilizes Llama 3, an open-source LLM from Meta, deployed on edge servers for on-site inference
* Runs on Supermicro IoT SuperServer SYS-111E FWTR hardware
* The edge server features 4th Gen Intel Xeon Scalable Processors with up to 32 cores
* Supports up to 2TB of DDR5 RAM for handling large language models
* Includes PCIe 5.0 expansion slots for additional AI accelerators
* Cloud Integration:
* Incorporates ChatGPT for enhanced performance when needed
* Used specifically for faster token generation and complex instruction processing
* Provides redundancy and additional processing capability for the system
### LLMOps Pipeline and Workflow
The system implements a sophisticated processing pipeline:
1. Voice Input Processing:
* Accepts verbal commands in 98 different languages
* Converts speech to text using specialized models
* Handles unstructured natural language input from warehouse workers
2. Language Processing:
* Primary processing through edge-deployed Llama 3 for cost-effective, low-latency operations
* Secondary processing through ChatGPT for enhanced performance when required
* Translation of natural language commands into structured AGV instructions
3. Command Execution:
* Converted instructions are sent to the Zippy family of AGVs
* System handles real-time communication between human operators and robots
* Provides feedback and confirmation of command execution
### Production Considerations and Optimizations
The implementation shows careful consideration of several key LLMOps aspects:
* Cost Optimization:
* Strategic use of open-source Llama 3 for primary processing to control costs
* Selective use of ChatGPT for specific high-performance requirements
* Edge deployment reduces cloud computing costs and latency
* Performance Management:
* High-performance edge server configuration ensures responsive local processing
* Network optimization through multiple connectivity options (1GbE and 10GbE ports)
* Scalable architecture supporting additional GPU acceleration as needed
* Reliability and Availability:
* Hybrid architecture provides redundancy and failover capabilities
* Edge deployment ensures operation even with limited internet connectivity
* Hardware specification allows for future expansion and upgrades
### Integration with Existing Systems
The LLM system is tightly integrated with Addverb's existing robotics infrastructure:
* Works with the complete Zippy product family, including various payload capacities from 6kg to 2,000kg
* Interfaces with existing warehouse management systems
* Supports grid-based navigation and obstacle detection systems
### Operational Impact and Results
While specific metrics are not provided in the source material, the system appears to deliver several key benefits:
* Reduced dependence on specialized maintenance engineers
* Decreased AGV downtime through faster problem resolution
* Improved accessibility through multi-language support
* Enhanced operational efficiency in warehouse operations
### Critical Analysis
The implementation demonstrates several strengths:
* Practical use of edge computing for LLM deployment
* Thoughtful hybrid architecture balancing cost and performance
* Strong consideration for scalability and future expansion
However, there are several aspects that warrant consideration:
* The reliance on cloud-based ChatGPT could present challenges in environments with limited connectivity
* The system's performance metrics and error rates are not clearly documented
* The training and updating procedures for the language models are not detailed
### Future Considerations
The system appears well-positioned for future enhancements:
* Potential for additional language support
* Expansion of command capabilities
* Integration of newer language models as they become available
* Enhanced edge processing capabilities through hardware upgrades
This case study represents a practical example of LLMs being deployed in industrial settings, demonstrating how careful architecture and implementation choices can create effective production systems that solve real-world problems.
Start your new ML Project today with ZenML Pro
Join 1,000s of members already deploying models with ZenML.