Company
Doppel
Title
AI Agent Automation of Security Operations Center Analysis
Industry
Tech
Year
2025
Summary (short)
Doppel implemented an AI agent using OpenAI's o1 model to automate the analysis of potential security threats in their Security Operations Center (SOC). The system processes over 10 million websites, social media accounts, and mobile apps daily to identify phishing attacks. Through a combination of initial expert knowledge transfer and training on historical decisions, the AI agent achieved human-level performance, reducing SOC workloads by 30% within 30 days while maintaining lower false-positive rates than human analysts.
This case study examines how Doppel, a cybersecurity company specializing in digital risk protection, successfully deployed an AI agent to automate significant portions of their Security Operations Center (SOC) workload. The implementation demonstrates several key aspects of LLMOps in a high-stakes production environment where accuracy and reliability are crucial. The Problem Space and Context: Doppel operates in the challenging domain of cybersecurity, where they process and analyze over 10 million potential security threats daily across websites, social media accounts, and mobile apps. Their primary focus is on identifying and responding to phishing attacks. The traditional approach involved using basic machine learning to filter obvious false positives, followed by human analysts making nuanced decisions about threat legitimacy and appropriate responses. This process was time-consuming and scaled poorly with increasing threat volumes. Technical Implementation and LLMOps Approach: The company built their solution using OpenAI's o1 model, showcasing several sophisticated LLMOps practices: Knowledge Transfer and Training: * They developed a systematic approach to transfer human expert knowledge to the AI system, essentially teaching it the same way they train human analysts * The training covered multiple security domains including phishing, malware, and brand abuse * They augmented this with thousands of well-curated historical decisions, effectively capturing years of analyst expertise Production Implementation Considerations: * The system needed to handle unstructured data including screenshots and time-series activity * They implemented the ability to process customer-specific policies and generate explanations for decisions * The solution required integration with existing security infrastructure and workflows Continuous Learning and Feedback Loops: * They implemented a continuous learning system where the AI agent improves from new examples * This was especially crucial given the rapidly evolving nature of security threats * The feedback loop allows the system to adapt to new attack patterns quickly Performance Monitoring and Validation: * They established clear benchmarks comparing AI performance against human analysts * Metrics included false-positive rates and threat detection accuracy * The system's performance was monitored for both speed and accuracy of responses Key Technical Challenges and Solutions: One of the most significant challenges was developing an AI system capable of exercising judgment in complex scenarios. This required: * Processing multiple types of input data (text, images, time-series data) * Understanding and applying context-specific policies * Generating clear explanations for decisions * Maintaining consistent accuracy across edge cases The team approached these challenges through: * Careful curation of training data * Implementation of robust feedback mechanisms * Integration of domain-specific knowledge into the training process * Development of clear performance metrics and monitoring systems Results and Impact: The implementation proved highly successful with several key outcomes: * 30% reduction in SOC workload within 30 days * Lower false-positive rates compared to human analysts * Improved threat detection rates * Faster response times to potential threats * More efficient allocation of human analyst resources to complex cases Critical Analysis and Limitations: While the case study presents impressive results, it's important to note several considerations: * The specific details of how they validate the AI's decisions aren't fully explained * The long-term reliability and drift of the model isn't addressed * The extent of human oversight required isn't clearly detailed * The specific mechanisms for handling edge cases or high-risk decisions aren't fully described Future Implications and Scalability: The case study suggests significant potential for AI in cybersecurity operations, particularly in: * Automating routine decision-making processes * Scaling threat detection and response capabilities * Improving response times to emerging threats * Enabling more efficient use of human expertise The implementation demonstrates a sophisticated understanding of LLMOps principles, particularly in: * The careful approach to training and knowledge transfer * The implementation of continuous learning mechanisms * The integration with existing workflows and systems * The focus on measurable performance metrics This case study represents a significant example of LLMs being successfully deployed in a high-stakes production environment where errors could have serious consequences. The approach taken shows careful consideration of the challenges involved in deploying AI systems in security-critical applications, while also demonstrating the potential for significant efficiency gains when such systems are properly implemented.

Start your new ML Project today with ZenML Pro

Join 1,000s of members already deploying models with ZenML.