Company
Addverb
Title
Multi-Lingual Voice Control System for AGV Management Using Edge LLMs
Industry
Tech
Year
2024
Summary (short)
Addverb developed an AI-powered voice control system for AGV (Automated Guided Vehicle) maintenance that enables warehouse workers to communicate with robots in their native language. The system uses a combination of edge-deployed Llama 3 and cloud-based ChatGPT to translate natural language commands from 98 different languages into AGV instructions, significantly reducing maintenance downtime and improving operational efficiency.
Addverb, a robotics company founded in 2016, has implemented an innovative LLMOps solution to address the growing challenge of AGV fleet maintenance in warehouse environments. This case study demonstrates a practical application of large language models in industrial robotics, combining edge computing and cloud services to create a robust production system. ### System Overview and Business Context The company faced a significant challenge in warehouse operations where maintenance issues with Automated Guided Vehicles (AGVs) required specialized engineers, leading to potential delays and operational inefficiencies. To address this, they developed an AI-powered voice control system that allows any warehouse worker, regardless of technical expertise or language background, to communicate with and manage AGV maintenance tasks. ### Technical Architecture and LLM Implementation The solution employs a hybrid approach to LLM deployment, utilizing both edge and cloud computing: * Edge Deployment: * Utilizes Llama 3, an open-source LLM from Meta, deployed on edge servers for on-site inference * Runs on Supermicro IoT SuperServer SYS-111E FWTR hardware * The edge server features 4th Gen Intel Xeon Scalable Processors with up to 32 cores * Supports up to 2TB of DDR5 RAM for handling large language models * Includes PCIe 5.0 expansion slots for additional AI accelerators * Cloud Integration: * Incorporates ChatGPT for enhanced performance when needed * Used specifically for faster token generation and complex instruction processing * Provides redundancy and additional processing capability for the system ### LLMOps Pipeline and Workflow The system implements a sophisticated processing pipeline: 1. Voice Input Processing: * Accepts verbal commands in 98 different languages * Converts speech to text using specialized models * Handles unstructured natural language input from warehouse workers 2. Language Processing: * Primary processing through edge-deployed Llama 3 for cost-effective, low-latency operations * Secondary processing through ChatGPT for enhanced performance when required * Translation of natural language commands into structured AGV instructions 3. Command Execution: * Converted instructions are sent to the Zippy family of AGVs * System handles real-time communication between human operators and robots * Provides feedback and confirmation of command execution ### Production Considerations and Optimizations The implementation shows careful consideration of several key LLMOps aspects: * Cost Optimization: * Strategic use of open-source Llama 3 for primary processing to control costs * Selective use of ChatGPT for specific high-performance requirements * Edge deployment reduces cloud computing costs and latency * Performance Management: * High-performance edge server configuration ensures responsive local processing * Network optimization through multiple connectivity options (1GbE and 10GbE ports) * Scalable architecture supporting additional GPU acceleration as needed * Reliability and Availability: * Hybrid architecture provides redundancy and failover capabilities * Edge deployment ensures operation even with limited internet connectivity * Hardware specification allows for future expansion and upgrades ### Integration with Existing Systems The LLM system is tightly integrated with Addverb's existing robotics infrastructure: * Works with the complete Zippy product family, including various payload capacities from 6kg to 2,000kg * Interfaces with existing warehouse management systems * Supports grid-based navigation and obstacle detection systems ### Operational Impact and Results While specific metrics are not provided in the source material, the system appears to deliver several key benefits: * Reduced dependence on specialized maintenance engineers * Decreased AGV downtime through faster problem resolution * Improved accessibility through multi-language support * Enhanced operational efficiency in warehouse operations ### Critical Analysis The implementation demonstrates several strengths: * Practical use of edge computing for LLM deployment * Thoughtful hybrid architecture balancing cost and performance * Strong consideration for scalability and future expansion However, there are several aspects that warrant consideration: * The reliance on cloud-based ChatGPT could present challenges in environments with limited connectivity * The system's performance metrics and error rates are not clearly documented * The training and updating procedures for the language models are not detailed ### Future Considerations The system appears well-positioned for future enhancements: * Potential for additional language support * Expansion of command capabilities * Integration of newer language models as they become available * Enhanced edge processing capabilities through hardware upgrades This case study represents a practical example of LLMs being deployed in industrial settings, demonstrating how careful architecture and implementation choices can create effective production systems that solve real-world problems.

Start your new ML Project today with ZenML Pro

Join 1,000s of members already deploying models with ZenML.