Writer, an enterprise AI company founded in 2020, has evolved from building basic transformer models to delivering full-stack GenAI solutions for Fortune 500 companies. They've developed a comprehensive approach to enterprise LLM deployment that includes their own Palmera model series, graph-based RAG systems, and innovative self-evolving models. Their platform focuses on workflow automation and "action AI" in industries like healthcare and financial services, achieving significant efficiency gains through a hybrid approach that combines both no-code interfaces for business users and developer tools for IT teams.
Writer represents an interesting case study in the evolution of enterprise LLM deployment, showcasing how LLMOps practices and approaches have matured since 2020. The company's journey mirrors the broader industry transition from basic transformer models to sophisticated production AI systems.
# Company Background and Evolution
Writer began in 2020, during what they call "the old days" of transformers, even before GPT-2. They started by building statistical models and their first large language model of 128 million parameters, which took six months to develop. This evolved into their Palmera model series, which they now offer as part of a full-stack enterprise AI solution.
The company's evolution reflects three distinct phases in enterprise AI adoption:
* Pre-ChatGPT (2020-2022): Focusing on basic text generation and educating customers about generative AI
* Post-ChatGPT Initial Phase: Enterprises experimenting with building in-house models
* Current Phase (2024): Focus on practical value delivery and stable, reliable systems
# Technical Architecture and Implementation
Writer's technical approach combines several key components:
## RAG Implementation Evolution
Their RAG (Retrieval Augmented Generation) system has gone through multiple iterations:
* Started with basic full-text search using Solr and Elasticsearch
* Moved to semantic search
* Evolved to a graph-based system
* Currently uses a hybrid approach with:
* Postgres for graph storage (avoiding traditional graph databases for scalability)
* Fusion Decoder (based on Meta's research) for retrieval
* Dense space implementation for similarity matching
They notably moved away from traditional vector databases due to scalability challenges at enterprise scale, where systems need to handle millions of documents.
## Model Deployment and Integration
The platform supports multiple deployment options:
* On-premises deployment
* Cloud deployment
* Various hybrid configurations based on data security requirements
* Integration with existing enterprise systems
## Monitoring and Observability
Their platform includes comprehensive monitoring capabilities:
* Input/output tracking
* Performance metrics
* Audit logs
* Integration with tools like Langsmith for request monitoring
# Enterprise Integration Approach
Writer's approach to enterprise integration is particularly noteworthy for its hybrid methodology:
## Business User Empowerment
* No-code interface for workflow definition
* Natural language description of process steps
* Visual flow representation
* Built-in monitoring tools for business users
## IT Team Integration
* Code interface for custom implementations
* Library integration capabilities
* Security configuration options
* Performance optimization tools
## Workflow Implementation
They focus on "action AI" and workflow automation, with specific emphasis on:
* Healthcare claim processing (reducing 100-step workflows to 50-30 steps)
* Financial services applications (wealth management, portfolio risk management)
* Real-time analysis and reporting
# Evaluation and Performance Measurement
Writer has developed a sophisticated approach to model evaluation and performance measurement:
## Model Evaluation
* Created internal evaluation frameworks due to data contamination concerns
* Focus on enterprise-specific metrics rather than standard benchmarks
* Emphasis on function calling and hallucination measurement
* Human-in-the-loop evaluation (after finding automated evaluation using LLMs unreliable)
## Business Impact Measurement
* Focus on time savings and productivity metrics
* Integration with existing enterprise KPIs
* Customer-specific measurement frameworks
* Target of 99-100% accuracy for enterprise use cases
# Latest Innovation: Self-Evolving Models
Writer's newest development is their self-evolving model system:
* Models that learn and adapt in real-time without external systems
* Built-in active learning capabilities
* Self-reflection mechanisms for error correction
* Focus on improving tool calling and multi-step workflow accuracy
* Aims to achieve 99%+ accuracy in production
# Enterprise Adoption Patterns
The case study reveals interesting patterns in enterprise AI adoption:
* Movement away from pure "build vs. buy" to hybrid approaches
* Increasing focus on practical value over specific model capabilities
* Strong preference for transparent, controllable systems
* Growing importance of business user involvement in AI system development
* Partnership model between IT and business units for successful deployment
# Challenges and Lessons Learned
Key challenges and solutions identified include:
* RAG system scalability issues at enterprise scale
* Data contamination in model evaluation
* Need for custom evaluation frameworks
* Importance of human oversight in specific areas
* Balance between automation and control
This case study demonstrates the complexity of enterprise LLM deployment and the importance of a comprehensive approach that combines technical sophistication with practical business value. Writer's evolution from basic transformer models to full-stack enterprise AI solutions provides valuable insights into successful LLMOps practices at scale.
Start your new ML Project today with ZenML Pro
Join 1,000s of members already deploying models with ZenML.