Wealthsimple developed a comprehensive LLM platform to enable secure and productive use of generative AI across their organization. They started with a basic gateway for audit trails, evolved to include PII redaction, self-hosted models, and RAG capabilities, while focusing on user adoption and security. The platform now serves over half the company with 2,200+ daily messages, demonstrating successful enterprise-wide GenAI adoption while maintaining data security.
This case study examines how Wealthsimple, a Canadian financial services platform, implemented a comprehensive LLM platform to enable secure and productive use of generative AI across their organization. The journey spans from late 2022 through 2024, showcasing the evolution from initial experimentation to mature enterprise deployment.
### Initial Platform Development and Security Focus
The company's LLM journey began with addressing the fundamental challenge faced by many enterprises after ChatGPT's release: how to enable AI usage while protecting sensitive data. Their first step was building an LLM gateway that provided:
- Comprehensive audit trails of all LLM interactions
- Centralized access to multiple LLM providers
- Built-in retry and fallback mechanisms for reliability
- Free usage for employees (company-covered API costs)
To drive adoption, they implemented both incentives and gentle nudges. The platform was made free for employees, offered centralized access to multiple providers, and included improved reliability features. They used Slack notifications to remind users about the gateway when accessing external LLM services directly.
### Security Enhancements and Platform Evolution
Security remained a top priority, leading to several key developments:
1. PII Redaction Model: They developed and open-sourced their own PII detection and redaction model to prevent sensitive data leakage. However, this introduced some user experience challenges as the redaction sometimes affected answer relevancy.
2. Self-hosted Models: To address the limitations of PII redaction while maintaining security, they implemented self-hosted open source models using llama.cpp. This allowed unrestricted data usage within their secure cloud environment.
3. RAG Implementation: They introduced semantic search capabilities as their first RAG API, though initial adoption was lower than expected due to user experience challenges.
### Developer Experience and Tool Integration
The platform evolved to support better developer experience through:
- A data applications platform built with Python and Streamlit for rapid prototyping
- Development of "Boosterpack," a personal assistant tool grounded in Wealthsimple context
- Support for document upload and knowledge base creation
- Integration of reference links for fact-checking and source verification
### 2024 Developments and Industry Alignment
The platform continued to mature in 2024 with several significant developments:
1. API Standardization: They redesigned their API to align with OpenAI's specifications, recognizing the importance of compatibility with emerging industry standards and tools like LangChain.
2. Multimodal Support: Added capabilities for handling images and PDFs, which saw rapid adoption with nearly one-third of users leveraging these features weekly.
3. Cloud Integration: Implemented Amazon Bedrock integration, showing a shift in build-vs-buy strategy toward leveraging established cloud services while maintaining security and privacy requirements.
### Lessons Learned and Best Practices
Several key insights emerged from their implementation:
1. Usage Patterns:
- Programming support dominated usage (approximately 50%)
- Content generation/augmentation was the second most common use case
- Information retrieval and research formed the third major category
2. Tool Integration:
- Tools are most valuable when integrated into existing workflows
- Multiple separate tools create friction and reduce adoption
- Centralization of functionality is crucial for user adoption
3. Implementation Strategy:
- Security and privacy should be foundational considerations
- User behavior often differs from stated preferences
- Being deliberate about tool development and business alignment is crucial
### Results and Impact
The platform has achieved significant adoption metrics:
- Over 2,200 daily messages
- Nearly one-third of employees are weekly active users
- Over half of the company uses the platform monthly
### Technical Architecture Considerations
The implementation demonstrates several important architectural patterns:
1. Gateway Pattern: Centralized access and control point for all LLM interactions
2. Security-First Design: Built-in PII protection and audit capabilities
3. Hybrid Deployment: Mix of self-hosted and cloud-based models
4. Platform Approach: Providing building blocks for other teams to develop solutions
### Future Directions
The platform continues to evolve with focus areas including:
- Further integration with workflow tools
- Enhanced multimodal capabilities
- Improved RAG implementations
- Expanded self-hosted model options
The case study demonstrates how enterprises can successfully implement GenAI while maintaining security and driving adoption through careful attention to user needs and behavior patterns. It also shows the importance of being adaptable to changing requirements and technology landscapes while maintaining a focus on practical business value.
Start your new ML Project today with ZenML Pro
Join 1,000s of members already deploying models with ZenML.