BLOG
How to Build an AI-Enabled SOC: Lessons From Teams That Did It Without Ripping and Replacing

Somewhere in your stack, there's probably an AI feature you're paying for but not using. Maybe it's an ML-based alert scorer that nobody trusts. Maybe it's an automated triage workflow that got turned off after it miscategorized a real incident. Maybe it's a copilot that summarizes alerts but doesn't connect to your actual detection logic.
You're not alone. Industry data shows 40% of SOC teams use AI or ML tools without making them a defined part of operations, and 42% rely on out-of-the-box AI with zero customization. The tools got purchased. They didn't get operationalized.
The teams that actually made AI productive in their SOCs didn't start with procurement. They started with foundations: normalizing data, moving detection rules into version control, and documenting workflows before handing anything to a model. This article breaks down what those teams did differently, from structured telemetry and detection-as-code to phased rollout and trust-building, so you can turn shelfware into something your analysts actually rely on.
Key Takeaways
AI requires operational excellence as a prerequisite; it doesn't create it. Teams that documented workflows and normalized data before evaluating AI tools were better positioned to operationalize them, while teams that skipped foundations ended up with purchased-but-unproductive platforms.
Detection-as-code is the starting point, not an advanced practice. Python-based rules in version control, with testing, give AI agents a clean foundation to amplify. Without structured data and quality detection logic, AI amplifies false positives instead of signal.
Start with alert triage, then expand. A phased adoption framework, from shadow observation to supervised augmentation to scaled automation, builds trust gradually. Teams that skip critical validation steps create adoption problems later.
The metrics that matter go beyond MTTD and MTTR. Automation coverage rate, analyst-to-alert ratio, false positive trends, and threat coverage growth tell the real story of whether AI is actually changing how your team operates.
Why Most AI SOC Initiatives Stall Before They Start
Most stalled AI SOC efforts fail before rollout because the blockers are operational, not technical. Here's how premature buying creates shelfware, and what the teams that succeeded did before procurement.
The Tool Trap: Buying AI Before Your Team Is Ready
Most SOCs deploy AI tools without operationalizing them. Recent survey data reveals a familiar pattern:
69% of SOCs still handle metrics reporting manually or mostly manually.
40% use AI or ML tools without integrating them into defined operations.
42% rely on out-of-the-box AI or ML with no customization.
The pattern is consistent: adoption starts at procurement instead of workflow design.
The underlying problem is rarely the AI itself. 54% of organizations cite insufficient knowledge and skills as the primary barrier to AI adoption in cybersecurity. If your playbooks aren't written down and your alert triage process isn't consistent, AI will amplify your inconsistency, not fix it.
What Teams That Succeeded Did Differently
Successful small security teams treated AI adoption as a 6 to 9 month operational shift, not a technology purchase. The critical differentiator was preparation before procurement.
According to Shawn Chakravarty, who says he built SOCs at Upwork, PayPal, and American Express, the Upwork team documented existing workflows first, planned in small chunks rather than automating everything at once, and adopted a team norm: "If you do something more than three times, it is time to automate."
One useful adoption framework breaks teams into three categories:
Takers, who use AI tools as delivered.
Shapers, who adjust tools to fit workflows.
Makers, who build custom AI solutions.
For lean teams of three to ten people, start as a Taker, but with operational integration into documented workflows, not standalone deployment.
Get Your Data Foundation Right First
AI depends on two foundations to produce trustworthy output: structured telemetry and detection-as-code. Get these wrong, and you're feeding an expensive model garbage data and expecting clean output.
Why Structured Data Is the Prerequisite for Trustworthy AI
Structured data and detection-as-code determine whether AI amplifies quality signal or amplifies false positives. This is the pattern we see consistently across Panther deployments. Three data qualities matter most before you layer on AI:
Consistency across sources: Endpoints, networks, SaaS apps, and cloud telemetry need common field names and data types.
Completeness for context: Missing user, asset, or process context forces AI to guess during triage.
Normalization at ingest: Applying schema early makes downstream reasoning more reliable than waiting for post-processing.
The key architectural distinction is structured-at-ingest versus post-processing normalization. Schema frameworks like OCSF normalize raw logs into consistent field names, data types, and semantic meaning, so AI agents can reliably access structured fields like actor, device, and process without parsing variations across log sources.
That consistency is what lets AI move from summarizing alerts to reasoning across them.
Detection-as-Code as the Starting Point
Detection-as-code transforms security rules from manually configured SIEM content into version-controlled, tested, CI/CD-deployed software artifacts. This is the foundation that makes AI augmentation trustworthy.
A mature detection workflow means rules live in Git, get peer-reviewed via pull requests, pass automated unit tests, and deploy through CI/CD pipelines. Here's a simplified example of what that looks like in practice:
The value is not just the Python code, but the fact that the logic is visible, testable, and reviewable before deployment.
This played out in Wolt's experience, where the security team adopted an everything-as-code approach with Git integration, CI/CD pipelines, and integrated testing on a cloud-native SIEM. As their Security Operations Lead Anssi Matti Helin put it: "Making changes with the integrated testing in Panther Analysis Tool makes it possible to be much more fearless with the changes. I don't need to wonder if this will break something, because I know if it will break something before I even deploy it."
In Panther, detection rules are written in Python and deployed through CI/CD pipelines. AI can assist with generating initial detection logic from natural language descriptions using AI Detection Builder, but the output is fully visible code that your team reviews, tests, and owns.
Redesign Roles Before You Redesign Workflows
AI adoption changes job design as much as process design. If you skip the people side, even a technically sound rollout will stall.
From Alert Operators to Threat Hunters: The Skill Shift
SOC analyst roles are shifting from reactive alert operators to proactive threat hunters as AI takes over routine Tier-1 triage. Workforce research tracks this trend at the industry level. This shift isn't optional: manual operations cannot sustain current alert volume. We see it play out in how teams use Panther: less time closing false positives, more time writing detection rules and hunting threats.
48% of cybersecurity professionals feel exhausted from trying to stay current on threats, while 55% of SOC respondents have considered walking away from their jobs due to pressure. SANS updated its SEC450 curriculum in Summer 2025 to expand coverage of detection engineering, threat-informed defense, cloud logging, and AI integration, signaling these are increasingly essential blue-team capabilities.
What Your Team Needs to Learn (and What They Can Stop Doing)
The gap isn't headcount; it's capabilities. 52% of organizations cite "not having the right staff" as their biggest concern, not insufficient headcount. The practical implication: retraining matters more than adding seats.
What to develop:
AI collaboration and prompt engineering: Directing and validating AI systems.
Detection engineering: Designing and optimizing detection logic.
Automation workflow design: Configuring playbooks and defining escalation logic.
What to stop doing manually:
Alert triage and initial enrichment
Repetitive data collection and log correlation
False positive investigation, which consumes a large share of current workload in alert-heavy SOCs
This is the real role redesign: less copy-paste investigation work, more judgment, tuning, and coverage expansion.
Phase AI In Without Ripping and Replacing
A phased rollout works best when each step earns the next one. Start with the use case that offers the clearest operational relief, then put guardrails around it so trust grows alongside automation.
Start With Alert Triage, Then Expand
Alert triage is the best starting point because it addresses the highest-volume workload, presents the lowest risk, and provides a natural progression path. Nearly half of all alerts go completely uninvestigated and two-thirds of SOC teams cannot keep pace.
Phase 1: Shadow observation (months 0 to 3). Deploy AI in parallel with human analysts. AI observes and classifies but takes no action while you track accuracy and establish baselines.
Phase 2: Supervised augmentation (months 3 to 6). Enable AI-powered alert enrichment and prioritization while maintaining human approval for all response actions and creating feedback loops for corrections.
Phase 3: Scaled automation (months 6 to 9+). Automate routine triage for well-understood alert types while preserving human decision authority for complex or novel threats.
Cresta's team followed this progression with Panther AI, achieving at least 50% faster triage, especially in complex investigations. As their Head of Security Robert Kugler noted: "Panther AI cuts triage time by at least 50%, especially in more complex investigations."
Critically, the platform shows its work: enrichments, detection logic, and evidence are all visible, not hidden behind a black box.
Build Trust Through Transparency, Not Blind Automation
Trust limits AI adoption speed more than model quality does. The NIST AI RMF includes transparency as a characteristic of trustworthy AI, stating directly: "Perceptions of negative risk stem from a lack of ability to make sense of, or contextualize, system output appropriately."
In practice, trustworthy AI in the SOC should make three things obvious:
What the system did: The alert changes, enrichment steps, and actions taken.
Why it did it: The evidence, detection logic, and reasoning path behind the recommendation.
Where humans stay in control: Approval points, escalation boundaries, and audit logs.
This is the principle behind human-in-the-loop design: AI pauses before executing sensitive actions, surfaces its reasoning for review, and logs every decision in an audit trail. When every alert escalation shows what happened, how AI reached its conclusion, and the evidence collected, analysts can validate outcomes rather than blindly trusting them.
Measure What Actually Changed
The point is not to collect more metrics. It's to choose metrics that reveal whether human work actually changed.
The Metrics That Matter Beyond MTTD and MTTR
Choosing the wrong metrics can actively harm your SOC. If your dashboards track activity instead of outcomes, your team optimizes for closing tickets rather than improving detection. Effective metrics need to tie directly to your organization's mission and security goals.
Operational efficiency:
Automation coverage rate: Percentage of tasks handled by AI workflows versus manual analyst work.
False positive rate trend: Percentage of alerts that prove to be false positives over time; every false positive eliminated is investigation time recovered.
Alert triage time: Time from alert creation to triage decision, the direct measure of AI's acceleration effect.
Team sustainability:
Analyst-to-alert ratio expansion: How many alerts each analyst effectively manages per shift over time, demonstrating capacity expansion without proportional headcount growth.
Detection capability:
Threat coverage growth: Breadth of MITRE ATT&CK techniques covered by active detection rules, tracked before and after AI adoption. If AI frees analyst time from triage, that time should translate into new detection rules and expanded coverage.
These metrics work together. If triage time drops but threat coverage stays flat, you may have reduced toil without reinvesting the time in better detection.
What AI-Enabled SOC Teams Look Like a Year In
Across documented AI SOC implementations, results often show workforce reallocation rather than immediate headcount reduction. Analysts shift from reactive responders to overseers of automated systems, spending time on threat hunting and detection engineering rather than triaging false positives.
A year in, common outcomes include:
More after-hours coverage through AI triage without adding headcount.
Cleaner detection logic as AI feedback loops identify noisy rules.
Less alert backlog as routine triage work gets compressed.
More threat coverage because analyst time shifts into writing and tuning detection rules.
The teams that reach this point share one characteristic: they treated AI as a force multiplier for an already-functioning operation, not a rescue plan for a broken one. They invested in structured data, adopted detection-as-code, defined governance before procurement, and phased AI in with transparency at every step.
Share:
RESOURCES






