From detection to resolution in minutes — our Agentic security operations platform intelligently investigate and remediate threats and keeps attackers one step behind
🌐 Visit Blogs.arambhlabs.com
🌐 Blogs.arambhlabs.com besuchen
Write rieview✍️ Rezension schreiben✍️ Get Badge!🏷️ Abzeichen holen!🏷️ Edit entry⚙️ Eintrag bearbeiten⚙️ News📰 Neuigkeiten📰
Tags: agentic attackers detection intelligently investigate minutes operations platform remediate resolution security threats
60313 Frankfurt am Main, DE Germany, EU Europe, latitude: 50.1169, longitude: 8.6837
We sat down with Kathy Del Gesso, a CISO who's been in security leadership for years, to talk about what's actually changing right now. The conversation went broader than we expected. CISOs aren't just securing systems anymore. They're trying to figure out how to govern AI agents that act autonomously, manage thousands of machine identities, and make decisions that happen faster than humans can track.
The security landscape is changing along three major axes.
Risk is concentrating in cloud runtime. This is where code, data, identities, and AI systems are actually executing. Static controls can't keep up anymore. The old model of securing perimeters and data at rest doesn't work when everything is dynamic and distributed.
Identity has become the new weak spot. We're not just talking about user accounts anymore. Organizations are dealing with thousands of machine identities, tokens, and AI agents making decisions automatically. The attack surface has expanded from people to include all these non-human entities operating at machine speed.
CISOs are being asked to shift from blockers to growth enablers. Executives want security teams to be part of building customer trust and winning deals, not just the department that says no. This is a fundamental reframing of the security function's relationship to the business.
When AI agents start taking actions in your environment, you're dealing with a new category of risk: unintended behavior at machine speed.
These aren't classic malicious attacks. They're accidental misuse happening very quickly. Agents chain APIs in unexpected ways. They escalate privileges through automation. They spin up workloads faster than you can monitor them.
The challenge isn't securing a network perimeter anymore. You're not thinking about controlled entry points or static data storage. There are multiple ways into cloud runtime, and traditional security models weren't built for this level of dynamism.
Kathy pointed out that agents will take actions we didn't explicitly approve. That's both the promise and the problem. The promise is automation and efficiency. The problem is losing visibility into what's actually happening and why.
The obvious threats are getting attention. API keys, service accounts, non-human identities, these are on most security teams' radar now.
What's less obvious is how unprepared SOCs are for this shift. Most teams haven't adapted their detection logic to identify these new patterns. They're still looking for traditional indicators of compromise, not unintended agent behavior or automated privilege escalation.
SOCs need new signals, better context, and updated playbooks. They're drowning in data from various tools, alerts, and logs. Correlating that context was already hard. Add in the non-human component, and it becomes exponentially harder.
The defender versus attacker dynamic is evolving into what Kathy called "Spy vs. Spy," the old cartoon where each side tries to outwit the other. Attackers are adapting fast. They're using AI for better phishing, faster reconnaissance, more personalized social engineering that's harder to spot.
But the real shift is toward the cloud control plane. API keys, service accounts, model endpoints. These are becoming the focus because traditional entry points are getting harder as organizations deploy AI-powered defense tools. So attackers are going after places where AI behaves unpredictably: model manipulation, prompt injection, influencing automated decision making.
AI is already helping with high-volume, repetitive work. Alert triage, noise reduction, log correlation, incident summaries. That first draft investigation report that sometimes doesn't happen as quickly as you'd like.
AI connects signals faster than humans can. It makes Tier 1 work more efficient. Some companies have fully automated Tier 1 with AI. A few advanced organizations have even automated Tier 2.
But humans are still needed for judgment calls. Understanding intent and business context. Making decisions about whether something is a real issue or not. Aggregating risk to determine if it needs to be addressed.
The conversation we had with Anton last week touched on the same point: alerts are a byproduct of detection rules. Unless you tie AI back to detection engineering, you're just processing garbage faster. The real opportunity is in adaptive defense. Proactive threat hunting in your environment. Writing adaptive detection rules based on what you're seeing. Building and testing those rules in simulated environments before deploying them.
There will always be a line between what AI can decide and what requires human judgment. High-stakes decisions stay with humans. Low-stakes decisions can potentially be automated. The nuance is in determining what counts as high-stakes versus low-stakes, and that's something you figure out working with the teams where this gets deployed.
AI isn't just everywhere in marketing speak. It's actually reshaping what CISOs do day to day.
Kathy mentioned the role is expanding from protecting systems to securing decision making. When AI agents can act across cloud environments, modifying infrastructure, spinning up workloads, moving data autonomously, you're no longer guarding perimeters. You're governing behavior.
The CISO's responsibility now extends to both human and machine actors. They're being pulled into product strategy, engineering decisions, and leadership conversations. Because AI isn't just another tool. It's part of the operating model.
This creates a hybrid role. You're still a security leader, but you're increasingly focused on risk management and AI governance. And as activity shifts to cloud runtime, new questions emerge about how traditional security roles interact with platform engineering and other functions in this new landscape.
Organizations might need to rearchitect their team structures to follow the technology. That's still being figured out.
AI coding tools have made it easier to build things in-house. Someone can spin up the equivalent of DocuSign in a weekend now. That changes the build versus buy equation, but not in the way people think.
There are a few factors to consider. First is speed versus depth. Buying gets you capabilities fast, which matters when you needed something yesterday. Building lets you dig deeper and tailor to your specific architecture and risk profile.
Then there's talent. Do you have the in-house expertise to build AI systems? More importantly, do you have the resources to maintain them long-term? Most teams underestimate the upkeep. The person who built it might not be there in a couple years. Tech has high turnover. Priorities change. The product needs to evolve over time to keep meeting your needs.
Organizations often misjudge this decision. They underestimate the long-term cost and complexity of building. They think of it as a one-time feature. But with AI, you need constant tuning, monitoring, retraining, and support. Then there's integration with your tech stack and the ongoing talent requirements.
The result of getting this wrong is projects that take longer, cost more, and deliver less value than buying a mature solution would have.
That said, if you have an expert team in a specific area and feel confident building something, it might make sense. Leave the things you don't have deep expertise in to people who do that work every day.
There's also a spectrum here. Companies like Google and Meta have the engineering power to build most things internally, and often prefer to. But most companies don't have that luxury. It really depends on company culture and risk tolerance.
The most underestimated shift is the level of autonomy AI systems will have in day-to-day operations.
AI agents won't just assist. They'll make changes to cloud environments, move data, deploy code, and interact across systems automatically. They'll interact with other agents. This shifts the entire risk model from human-driven to machine-driven.
The risk isn't just more attacks. It's also losing visibility into why an AI system acted a particular way. Not having the guardrails to control it.
Teams need to prepare for a new level of autonomy, oversight, and monitoring of AI systems. Some companies are more advanced than others, but as a whole, we're still figuring this out.
The Palo Alto acquisition of Chronosphere seems timely in this context. Observability of everything agents can do and potentially will do becomes critical. It's a missing piece that companies are scrambling to solve, and you're seeing consolidation around it. Larger companies want to stay competitive, and they're acquiring capabilities rather than building them
The goal isn't to resist these changes. The goal is to understand your unique environment, find the right tools for visibility, and recognize where you need to grow your security controls.
Nobody has a 100% foolproof answer yet. CISOs are learning from each other in forums and peer discussions. The field is evolving too fast for certainty. But recognizing where you're vulnerable and understanding what's in your control versus what cloud providers manage is a good start.
At Arambh Labs, we built specialized AI agent swarms for exactly this problem. Our platform covers alert triage, threat hunting, and adaptive detection engineering. The goal is simple: give SOC teams visibility and control as their environments fill up with autonomous AI systems. Security analysts shouldn't spend their time on repetitive work. They should focus on the high-stakes decisions that actually need human judgment. That's what we're building for.
Curious how CISOs are navigating AI governance? Watch the full conversation with Kathy on our YouTube channel.
If you're ready to see how specialized AI agents can handle runtime security challenges in your environment, visit our website and schedule a demo with our team.
3.12.2025 20:05Why Runtime Security Is the New PerimeterWe sat down with Anton Chuvakin, one of the most outspoken voices in security operations, to cut through the AI SOC hype. What we got was a reality check on why alert fatigue has plagued SOCs for 20 years, why the traditional Level 1/2/3 analyst model is broken, and what it actually takes to build security operations that work in 2025.
Anton doesn't mince words. He's watched companies build 2002-style SOCs in 2022, using blueprints from old whitepapers as if they were timeless fundamentals. And now he's watching vendors promise that AI will replace entire security teams. Spoiler: it won't.
The state of SOCs today is all over the map. On one end, you have modern detection and response operations built on engineering-first principles. These teams refuse to even call themselves SOCs because they don't think of themselves as operators, they're engineers. Think SRE or DevOps, but for security.
On the other end, you have organizations still running late 90s SOC models. Big monitors. Rigid shifts. Analysts sitting in chairs triaging alerts the same way they did two decades ago. They've got slightly better tools now, EDR instead of nothing, modern network detection instead of whatever they had back then. But the fundamental structure? Unchanged.
Some companies literally picked up a white paper from 2002 and built everything according to that blueprint. Nobody told them to check the date. They thought they were following fundamentals.
The problem is that when 30 AI SOC vendors show up promising to fix everything, most of them are trying to improve a 1970s NOC with 2025 technology. You can add AI to a broken model, but you're still working with a broken model.
Alert fatigue has been the top complaint in security operations since 2005. If you took a time machine back then and asked a SOC analyst what their biggest problem was, they'd say alert fatigue. Fast forward to 2025, and we're saying the exact same thing.
That staying power is what makes it worth exploring. Because if we've been complaining about the same thing for 20 years, maybe we're looking at a symptom instead of the root cause.
Anton pointed out the real problem: we suck at detection.
That's why watching vendors rush to build AI agents for alert triage feels like missing the point. Faster triage is handy, but if your underlying detection is broken, doing triage faster just means you're processing garbage more efficiently.
Alerts don't just happen like weather. They come from rules someone wrote. And in many SOCs, there's zero relationship between the people who write those rules and the analysts dealing with the alerts. When analysts have never met the content engineers building detection logic, when they never talk about what makes alerts fire, there's almost no hope of fixing the problem.
That bridge between detection engineering and SOC operations is where the real work happens. Whether you call it content authoring or detection engineering, it's the same idea. You need that connection.
Here's where things get contentious. The phrase "AI SOC" is misleading marketing.
AI in a SOC? Yes. AI SOC? No.
There are real, functioning use cases for AI and agents in security operations. The problem is the expectation gap that sometimes emerges. When vendors sell these tools, decision-makers sometimes interpret "AI SOC" to mean they can replace their entire human security team with automation.
That's the disconnect. It's where the technology promise gets misunderstood.
The issue isn't always vendor marketing. Sometimes it's organizational pressure. Security operations are expensive. Leaders looking at budgets want to know if technology can reduce headcount. It's a reasonable business question, but it's the wrong frame for this technology.
When someone hears "AI SOC," the risk is they think it means eliminating the security team rather than amplifying what that team can do. That's not how this technology works, and it's not what it should be sold as.
The value isn't replacement. It's transformation of how the work gets done.
Traditional SOCs organize analysts into levels. Level 1 does basic triage. Level 2 handles escalations. Level 3 takes the hardest cases. It's a model borrowed from NOCs in the 1990s.
The question isn't whether AI can automate Level 1 work. The question is whether that's even the right frame anymore.
Netflix wrote about sockless detection back in 2018. Other companies followed. The whole idea is to dissolve those rigid levels and focus on skills instead of tiers. Do engineering-first designs. Build systems where humans focus on what actually requires human judgment.
Adding AI to a Level 1/2/3 structure just creates weird mental gymnastics. If AI replaces Level 1, what do Level 2 analysts do? Get alerts from the AI? So now the bottom layer is machines feeding humans, but if the machines generate bad alerts, who do you complain to?
It's easier to add AI to a more modern detection and response model than to patch it onto a structure that was outdated 15 years ago. You're not adding AI to a 1990s SOC. You're adding AI to a 1970s NOC. If that doesn't scare you into rethinking your approach, nothing will.
If you had to build security operations from the ground up today, what would you do?
Anton mentioned he's working on a paper with Deloitte that describes exactly this. Start with engineering-first principles. Base it on what's being called Autonomic Security Operations or sockless detection. Not because you're trying to be trendy, but because it's designed for how threats and technology actually work now.
Some organizations have done this by tearing down their existing SOC and rebuilding. Not gradual improvements. Full teardown and rebuild. It takes 2-3 years of work, but it works.
From day one, you'd build something AI-ready. That means data quality suitable for AI. Modern pipelines. Process maturity that accounts for the fact that AI makes mistakes. You'd use AI from the starting point, but you wouldn't call it "AI native" because that term has been ruined by marketing.
The goal isn't to ask which chairs Level 2 analysts should use or how handoffs should work. It's to build something designed for the future.
In a properly built detection and response operation, humans focus on a few specific things:
Decide what to build. This is fundamentally human. Look at available resources, assess risks, decide what's worth detecting. AI can offer advice, but humans make the call. Architectural analysis and design stay human-led.
Oversee how machines function. Validate that pipelines work. Figure out why volume is dropping. Catch when business does something that makes detection less effective. Ask why detection isn't performing as well as it should. Humans do this with AI help, but it's still human-led.
Top-shelf threat hunting. Hunting is, by definition, a hypothesis-led process. Gen AI helps generate hypotheses based on threat intel. But certain types of hunting for sophisticated attackers will remain human with tools, not just asking a machine to handle it.
Judgment calls when machines stumble. And they will stumble, because attackers want them to. The adversary gets a vote. When machines are uncertain, do the wrong thing, or get stuck, humans step in. Nobody wants an agent that was supposed to finish in 2 hours still deliberating with itself the next morning.
Risk acceptance decisions. These are going to stay human for a very long time. It's one of those decisions where delegating to a machine creates massive grief.
A lot of vendors claim they've built threat hunting agents. What they've actually built is something that generates hypotheses or runs detections.
Today, hunting teams face long lists of things to try. You might allocate a day to persistence mechanisms, another day to exfiltration methods. Run through everything you can think of. But these lists have gotten huge, and environments are complex.
The collaboration model makes sense: machines run the hypotheses, bring in data, do preliminary analysis. They tell you that out of 30 persistence checks, 27 produced nothing. Two look a little suspicious. One looks probably bad.
Then humans take over. The "probably bad" one turns out to be your development team using a slightly non-standard approach to inject stuff in memory. It matches threat actor behavior, but it's not a threat actor doing it. The machine was right that it looked suspicious. It just needed a human to provide context.
The two interesting ones need deeper investigation. Run more queries. Check other things. Either you discover an attacker, or you confirm you're clean and move on.
This is collaboration. The human role is very much there.
You pay for AI SOC tools or build AI into your SOC because you expect something to be better. But if you don't measure what "better" means, how do you know?
A lot of people rush straight to speed metrics. MTTD, MTTR. They start chanting these abbreviations like mantras. Speed metrics are fine. You almost expect machines to be faster. But if you obsess about speed without looking at quality, effectiveness, or whether you actually achieved the result, you're going to lose.
It's much faster to just click "resolve" on every alert. You could write a dumb script to do it. Your speed metrics would look amazing. But you'd be 10 times worse at actually protecting anything.
If something is two times faster, two times cheaper, and 10 times worse, you're not saving money. Your speed metric looks good. Your cost metric looks good. But your actual security is in hell.
Balance speed metrics with coverage. Detection quality requires detection breadth. Are you detecting everything you need to detect? That means knowing what you need to detect first, then checking if you're actually catching it.
For critical threats, have multiple layers. EDR, logs, NDR, other tools. Then have a machine look at all of them and say, "Hey, I think NDR missed it and EDR missed it, but here in the log it's a little suspicious. Human, take a look."
That's a win for AI. It guided a human to the right spot.
The goal isn't to avoid AI or refuse to modernize. The goal is to be honest about what AI can and can't do, what humans should and shouldn't do, and what your security operations actually need.
If you have metrics you already track, adopting AI should make something measurably better. That sounds obvious, but it's not always how this plays out.
More importantly, don't let AI vendors sell you on replacing humans entirely. Don't let executives think they can fire the security team and run everything on automation. And definitely don't take a 2002 SOC model, bolt AI onto it, and expect transformation.
We believe the path forward starts from first principles. Build for engineering-led operations. Use AI where it makes sense. Keep humans in the loop for judgment, architecture, and the work that actually requires human intelligence. At Arambh Labs, we're building agentic AI that amplifies what security teams can do, not replaces them. The future of security operations isn't humanless. It's human intelligence freed from repetitive work and focused on what matters.
Want to hear the full conversation? Stream the complete episode on our YouTube channel.
Ready to see how agentic AI can transform your SOC? Visit our website and book a demo.
18.11.2025 19:42Why Your SOC Still Looks Like It's 2002 (And What AI Should Actually Fix)Based on interview with Sunil Mallik, Head of Cybersecurity Architecture and Engineering at PayPal and former CISO at Discover
We sat down with Sunil to talk about agentic AI in financial services. What we got was a masterclass in how security actually works when you're protecting billions of dollars in transactions across mainframes, cloud infrastructure, and everything in between.
Every financial services company has plans to adopt agentic AI. But the speed varies wildly, and risk appetite is the deciding factor. There are still fundamental challenges without clear answers, and the organizations moving fastest aren't necessarily the ones who'll get it right.
Before we get to AI, Sunil told us a story that reshaped how he thinks about security. His company decided to test remote work readiness. They sent everyone home on a Thursday for a practice run. "No one came back for a year and a half," he said.
That moment reshaped financial security. Teams had to enable remote work at scale while securing it, and they were figuring it out in real time. New tools like Miro went from novelty to necessity. The line between internal and external networks blurred so much that Sunil now says, "Everything is external."
Then, just as security teams caught their breath, the AI boom hit.
Here's what makes financial services security uniquely hard: you're protecting three eras of technology at once.
Some of the largest financial companies still run mainframes. They have on-premises data centers. They have modern cloud infrastructure. And increasingly, they rely on third parties managing critical pieces of their ecosystem.
The challenge isn't just covering all these environments. It's maintaining consistent security controls across systems that were built decades apart. When a transaction flows from cloud to data center to a third-party processor, how do you maintain context? How do you ensure your controls don't break when you patch something?
And here's the kicker: mapping out attack paths in this kind of environment is brutal. You need to see what an adversary sees from the outside and trace how they could exploit vulnerabilities across multiple ecosystems. That's how you design monitoring and implement controls that actually work.
Every financial services company we talked to has plans to adopt agentic AI. But the speed varies wildly, and risk appetite is the deciding factor.
"There are still risks we don't have complete answers on," we heard. Explainability and bias can't be solved with technical controls alone. They live in the messy intersection of technology and business process.
When financial institutions evaluate AI, they're weighing three things: cost, value delivered, and risk. And they have to answer these questions in an environment where regulators, board members, and customers expect serious due diligence.
The security challenges of agentic AI break down into three areas: identity, context, and action.
Identity gets complicated fast. An agent has its own non-human identity, but it also derives identity and entitlements from the human it's working for. The question is: how do you ensure an agent stays within the boundaries of what that human is allowed to do?
In a multi-agent system where agents hand off work to other agents, this becomes even harder. You need to carry that identity and those constraints through every handoff.
Context is about tying agent actions back to human intent. What is the agent doing? Why? Who authorized it? As context changes, you need continuous validation. This is where zero trust principles become critical, not as a tool but as a strategy.
Action is the big one. It's the most important challenge, and it gets philosophical: Is the agent acting in the best interest of the human it represents?
"You have to prove that. That's always been expected from technology we've deployed."
This means full auditability. Every action an agent takes needs to be logged so you can spot deviations from expected behavior.
Right now, SOC analysts spend most of their time in what gets called "noise versus news." They're drowning in alerts, triaging false positives, dealing with the same types of incidents over and over. It's Groundhog Day.
Agentic AI changes that equation. Analysts will spend less time on repetitive work and more time on things that matter: improving controls, writing new detection logic, finding blind spots in coverage.
The quality of SOC outputs will improve. You'll get better fidelity in analyst actions and better feedback loops to the rest of your security organization. Every incident becomes an opportunity to improve the controls that should have prevented it.
But humans aren't going anywhere. You'll still need analysts to validate critical actions like isolating users or network segments. The goal isn't full automation. It's freeing up human intelligence for work that actually requires it.
"There will always be a need for the analyst," we heard. The environment keeps changing. New technologies emerge. Business processes create noise. The idea that controls will ever work perfectly is fantasy.
The conversation turned candid when we asked about friction between security vendors and financial institutions.
Vendors often misunderstand how procurement works in large financial companies. A senior leader might love your product, but that doesn't mean you'll close the deal quickly. You still have to go through architecture review, third-party risk management, security oversight. The bigger the organization, the longer it takes.
These processes exist for a reason. They manage risk. They satisfy regulators and auditors. They might frustrate everyone involved, but they're not going away.
The other issue is communication. Vendors are rightfully proud of what they've built. But that outside-in view doesn't always translate clearly to the team you're pitching. Organizations differ in structure, priorities, and what they consider urgent.
In cybersecurity, there's no absolute assurance, only reasonable assurance. Every team has a backlog. They're working on the highest impact items. Your product might solve a real problem, but if it's not their top priority, it's not their top priority.
When we asked what advice security leaders should follow when integrating AI, the answer went back to basics.
Confidentiality, integrity, and availability. The CIA triad has been the foundation of cybersecurity for decades, and it still applies to AI. Most of the risks you've dealt with in other technologies apply here too.
But AI introduces unique risks that blur the line between technology and business. You need a different kind of shared responsibility model between technical teams and business owners.
Here's the non-negotiable part: when you integrate an agent into a customer-facing product, you're still responsible for ensuring that agent acts in the customer's best interest. The agent is working on their behalf. Their data needs to be protected. The integrity of their transactions needs to be maintained.
That's the foundation of customer trust, and you can't compromise on it.
There's also an amplification risk. If one human identity is compromised, and agents are acting on behalf of that identity, the problem propagates fast. It's not just one compromised account anymore. It's that account plus all the agent actions tied to it.
At Arambh Labs, we're building agentic AI specifically for security operations. This conversation reinforced something we already believed: the technology has to respect the fundamentals while solving the actual problems analysts face every day.
SOC teams don't need more alerts. They need to shift from noise to news. They need agents that understand identity, maintain context, and take actions that can be audited and validated. They need systems that amplify human intelligence instead of replacing it.
The future we discussed isn't about removing humans from the loop. It's about giving them back their time so they can do the work that actually requires human judgment.
Financial services will get there at different speeds based on their risk appetite and priorities. But the direction is clear. And the organizations that figure out how to balance innovation with customer trust will set the standard for everyone else.
Want to hear the full conversation? Stream the complete episode on our YouTube channel.
Ready to see how agentic AI can transform your SOC? Visit our website and book a demo.
4.11.2025 21:393 Security Problems every Financial Institution need to address in Agentic AI WorldThe landscape of digital identity has undergone a dramatic transformation over the past two decades. What began as simple username-password combinations has evolved into a sophisticated, critical component of cybersecurity infrastructure. In a recent conversation, identity security expert Mohit Vaish, CEO of CyberSolve shared insights with Neha Garg about how identity management has become the cornerstone of modern security architecture—and what the future holds as AI agents enter the equation.
According to Vaish, identity has progressed far beyond its original function as a backend authentication mechanism. "Identity has evolved significantly over the past two decades, moving from being merely IDs and passwords to a fundamental representation of individuals in the digital world," he explains.
This evolution can be understood in three key phases:
Phase 1: Basic Authentication – Identity served primarily as a gatekeeping function with simple credential verification.
Phase 2: Control Plane – Identity became a control mechanism for digital interactions, governing what users could access and do within systems.
Phase 3: Security Foundation – Today, identity has transformed into the core element of security architecture itself.
Garg emphasized the significance of this shift, noting how bold it is to position identity not as a component of security, but as security itself. This perspective fundamentally changes how organizations approach their security posture.
While the phrase "identity is the new perimeter" has become popular in cybersecurity circles, Vaish argues this terminology is misleading. Instead, he advocates for "identity-centric detection and response" as the more accurate framework.
Using an airport passport analogy, Vaish clarifies: "A passport enables trusted movement but is not the border itself. Similarly, identity facilitates secure interactions but is not the perimeter in isolation."
This distinction matters because it shifts focus from static boundary protection to dynamic threat detection and response centered on identity behavior.
Garg reinforces this perspective by highlighting a critical statistic: the majority of modern cyberattacks originate from compromised or stolen identities. This reality underscores why organizations need comprehensive identity detection and response strategies rather than treating identity as just another security layer.
One of the most persistent challenges facing security teams is the overwhelming volume of alerts and noise generated by identity monitoring systems. Vaish identifies the fundamental issue: organizations are conceptualizing identity incorrectly.
Vaish draws a compelling parallel to airport security evolution: "Modern systems detect identities first and then focus on their movement, which effectively reduces noise compared to older systems that focused on changes in pixels."
Traditional security approaches generate alerts based on every change or anomaly in system behavior—similar to monitoring every pixel change on a security camera. In contrast, identity-first approaches:
The key to reducing identity-related noise lies in fundamentally rethinking how organizations identify and track entities within their environments. Technologies that can accomplish this shift away from bolt-on approaches—where identity monitoring is simply added to existing security systems—to integrated, identity-native architectures.
As organizations increasingly deploy AI agents to act on behalf of humans, a new dimension of identity complexity emerges. Vaish explains that these agents create a multiplier effect on the attack surface while introducing unpredictable attack vectors.
Vaish proposes a comprehensive identity framework specifically designed for AI agents, which he calls "digital delegates." This model addresses several unique threat categories:
Agent Impersonation – Malicious actors creating fake agents or hijacking legitimate ones to bypass security controls.
Prompt Poisoning – Manipulating the instructions given to AI agents to alter their behavior in harmful ways.
Delegation Hijacks – Intercepting or redirecting the authority delegated from humans to their AI agents.
Cross-Agent Manipulation – Exploiting interactions between multiple agents to compromise systems laterally.
Garg highlights a particularly concerning aspect of AI agent security: inheritance of user capabilities. When large language models (LLMs) and AI agents inherit the full permissions of their human users, a single compromised user identity can have cascading effects throughout the entire agent network.
This reality makes managing and limiting the controls granted to AI agents not just important, but critical to organizational security.
Looking toward the future, Vaish envisions a convergence of detection, response, and identification into what he calls "identity-intent-driven defense." This approach represents a fundamental shift in how organizations think about identity security.
Continuous Trust Scoring – Real-time assessment of identity trustworthiness based on multiple factors and behaviors.
Behavior Baselining – Establishing normal patterns based on transactional activities rather than static rules.
Automated Response Loops – Systems capable of quarantining risky identities proactively, before breaches occur rather than responding after the fact.
Perhaps most significantly, Vaish suggests reframing "Identity and Access Management" (IAM) as "Intent and Action Management." This linguistic and conceptual shift emphasizes:
This evolution reflects a more dynamic, context-aware approach to identity security that aligns with the complexity of modern digital environments.
When asked how organizations should prepare their IAM systems for AI agents, Vaish recommends analyzing these entities through two complementary lenses:
AI Agents as Quasi-Humans:
AI Agents as Software Systems:
Vaish confirms that organizations can and should assign proficiency levels and roles to AI agents, similar to human user management. This approach enables:
As identity continues its evolution from simple authentication to the foundation of security architecture, organizations should consider:
The evolution of digital identity from simple credentials to the cornerstone of cybersecurity represents one of the most significant shifts in information security. As Mohit Vaish's insights reveal, this transformation is far from complete. With AI agents introducing new complexities and attack surfaces, the future of identity management lies in intent-driven defense, continuous trust assessment, and behavioral understanding.
Organizations that grasp this evolution and prepare accordingly will be better positioned to secure their digital environments in an increasingly complex threat landscape. The question is no longer whether identity is important to security—it's how quickly organizations can embrace identity as security itself.
22.10.2025 21:27Identity in the Agentic AI ageWe had a unique opportunity to sit down with Anant Thangaraju, who's a Field CISO at E+, to get his thoughts on the current landscape of agentic AI in cybersecurity. This blog post is based on the discussion between Neha Garg, Arambh Labs' CEO, and Anand Thangaraju. This blogpost covers Anand's views on modern CISO challenges, board communication, and the future of AI in security operations.
In today’s rapidly changing cybersecurity landscape, Chief Information Security Officers (CISOs) (chief information security officer) face unprecedented challenges. From navigating complex threat environments to justifying security investments to the board, the role has evolved far beyond traditional IT security management.
Field CISOs like Anand Thangaraju at EPlus are at the forefront of helping enterprises tackle these challenges. The modern chief information security officer is responsible for developing a comprehensive cyber strategy and robust risk management practices. Unlike traditional consulting roles, Field CISOs work directly with practicing CISOs, drawing from real-world operational experience in regulated environments to provide practical, actionable guidance.
One of the most critical mistakes organizations make is starting their security strategy with vendor selection. According to Thangaraju, successful cybersecurity programs begin with a fundamental question: What are our business priorities?
This business-first approach involves:
The hope for many security leaders is that agentic AI will provide the necessary orchestration layer to tie together existing security infrastructure and optimize security workflows more effectively. This could potentially solve the common problem of security tool sprawl while improving overall program efficiency.
Agentic AI systems represent a new generation of ai systems capable of autonomous decision-making and orchestration. Unlike traditional AI, these agentic ai systems can independently reason, plan, and act, enabling more dynamic and adaptive security operations.
Traditional cybersecurity metrics often fail to resonate with executive leadership. Metrics like “threats defended” or “automation hours saved” don’t translate effectively to business impact. Instead, successful CISOs are adopting new communication strategies that help business leaders gain insights into how security initiatives support organizational objectives.
Thangaraju advocates for a "vital signs" methodology when reporting to the board:
Effective board communication requires connecting cybersecurity initiatives to broader business transformations:
The cybersecurity industry has moved from the initial “Gen AI hype” into what Thangaraju calls the “agentic AI hype.” This new phase brings both opportunities and challenges:
Current Assumptions:
Reality Check:
The adoption pattern for agentic AI mirrors previous automation trends like Robotic Process Automation (RPA). Organizations will likely need to experiment to find the optimal balance, similar to how companies previously struggled to achieve board-mandated productivity gains from RPA implementations.
To build trust and demonstrate value, security organizations should begin their agentic AI journey with real world examples that show how agentic AI can analyze log data and enhance threat intelligence:
One of the most promising applications for agentic AI in security is the personal assistant or co-pilot model. This approach focuses on:
For broad, common use cases, organizations can:
Continuous improvement in data protection and efforts to protect sensitive data are essential for maintaining customer trust and preventing data leakage.
When developing agentic AI implementations, successful organizations focus on four key personas:
Successful agentic AI adoption requires a collaborative approach with strategic partners across the organization.
While industries like autonomous vehicles have made significant progress with complex, high-stakes AI implementations, the cybersecurity sector has been slower to adopt comprehensive platform approaches. This lag stems from:
Despite these challenges, the cybersecurity industry is positioned for significant advancement through thoughtful agentic AI adoption. Success requires:
The future of cybersecurity lies in the thoughtful integration of agentic AI technologies with human expertise. Field CISOs play a crucial role in this transformation, helping organizations navigate the gap between technological possibility and practical implementation.
Key takeaways for security leaders:
As the cybersecurity landscape continues to evolve, the organizations that succeed will be those that balance innovation with practicality, leveraging the power of agentic AI while maintaining the strategic thinking and contextual understanding that only human security
1.10.2025 19:30Field CISO Insights: How to Navigate Agentic AI in Cybersecurity and Align Security with Business GoalsIn today’s hyper-connected digital landscape, Security Operations Centers (SOCs) face an unprecedented challenge: SOC teams drown in alerts; response time is the difference between a minor incident and a catastrophic breach. This overwhelming volume of alerts leads to alert fatigue, a critical challenge where analysts become desensitized or overwhelmed, increasing the risk of missed threats and slower responses. With cyber attacks increasing in frequency and sophistication, the ability to detect, investigate, and respond to threats quickly has become the ultimate measure of cybersecurity effectiveness.
Mean Time to Respond (MTTR) in the SOC context represents the average duration from when a security alert is generated to when appropriate remediation actions are taken. This metric has evolved beyond a simple performance indicator—it’s now a critical business differentiator that directly impacts an organization’s security posture, compliance status, and bottom line. Achieving comprehensive threat detection is a key objective for modern SOCs. Arambh Labs ensures complete visibility and effective response across diverse environments.
Traditional SOCs struggle with MTTR measured in hours or even days. However, Arambh Labs' AI agents with advanced AI capabilities is revolutionizing this landscape, transforming response times from hours to minutes and fundamentally reshaping how security teams operate. We are focusing on creating intelligent, adaptive defense mechanisms that can keep pace with modern cyber threats.
Modern Security Operations Centers face a perfect storm of challenges that significantly impact their Mean Time to Respond: manual effort slows down security operations processes, leading to delays in threat detection and response.
Traditional SOC workflows often rely on static detection rules, which require frequent manual updates and can struggle to adapt to evolving threats. These limitations can result in missed threats or slow response times.
Outdated tools and approaches, such as legacy SOAR solutions, are resource-intensive to manage, difficult to integrate, and often fail to address the demands of modern cybersecurity environments.
Enterprise SOCs typically process millions of security signals daily, with many organizations reporting alert volumes exceeding 10,000 per day. Research indicates that up to 80% of these alerts are false positives, creating a needle-in-a-haystack scenario where critical threats can be buried under routine notifications. This overwhelming volume forces analysts to spend precious time sifting through noise instead of focusing on genuine threats.
The traditional SOC workflow relies heavily on human expertise for three critical phases: alert triage, context enrichment, and decision-making. Each of these stages can consume significant analyst hours, and often require a thorough human review to validate and respond to security alerts. This human review step, while essential for reducing false positives and ensuring accurate threat validation, can further delay response times:
The global cybersecurity workforce shortage exceeds 3.5 million professionals, with SOC analyst positions being particularly difficult to fill. This shortage means existing teams are overworked, leading to longer response cycles, increased burnout, and higher turnover rates. The result is a vicious cycle where fewer analysts must handle increasing alert volumes, further extending MTTR.
Extended MTTR carries severe consequences:
AI agents represent a paradigm shift in cybersecurity automation—they are autonomous, task-oriented artificial intelligence entities capable of investigating, enriching, and acting upon security incidents with minimal human intervention. Among these, AI SOC agents are specialized AI-driven tools designed specifically for Security Operations Centers (SOCs) to automate, enhance, and accelerate threat detection, investigation, and incident response. In this context, the roles of AI SOC analyst and AI SOC analysts refer to AI-powered agents that augment human analysts by automating triage, investigation, and response tasks, thereby improving efficiency and allowing human experts to focus on higher-value activities.
The broader ecosystem includes advanced AI tools that support SOC operations, while underlying AI systems provide the frameworks for coordination, dynamic task management, and learning among agents. A multi agent system enables multiple specialized agents to work together for advanced, coordinated security automation. Within modern security operations, both SOC agents (AI-powered automation tools) and SOC analysts (human experts supported by AI) are key components, working together to enhance efficiency and reduce response times. Unlike traditional rule-based systems, AI agents can adapt, learn, and make intelligent decisions based on context and experience.
While Security Orchestration, Automation, and Response (SOAR) platforms have provided valuable automation capabilities, they operate on pre-coded, deterministic workflows. However, legacy SOAR solutions often rely on static playbooks, are resource-intensive to manage, and struggle to adapt to the evolving demands of modern cybersecurity environments.
AI agents offer several advantages and are driving the broader trend of SOC automation:
AI agents can instantly analyze incoming alerts, cross-reference threat intelligence, and assign priority scores based on potential impact, likelihood of success, and business criticality.
These systems can automatically gather relevant information from multiple sources including:
During context enrichment and automated investigation, large language models and large language models (LLMs) are used to analyze security data, generate insights, and enhance decision-making by automating the interpretation of complex information.
Advanced AI agents can execute containment and remediation actions such as:
AI agents determine when human expertise is truly needed, ensuring analysts focus on complex scenarios that require human creativity and strategic thinking.
Selecting the right AI SOC platform is a critical decision that can have a lasting impact on your organization’s security posture. When evaluating potential solutions, it’s important to consider how well the platform integrates with your existing security tools and systems, as seamless interoperability is key to maximizing the value of your current investments.
Assess the platform’s level of automation and machine learning capabilities, as well as its support for multi-agent systems and coordinated response. A user-friendly interface and intuitive user experience can significantly enhance analyst productivity and reduce the learning curve. Scalability and flexibility are also essential, ensuring the platform can grow with your organization’s needs.
Consider the quality of customer support and training provided, as well as the total cost of ownership and expected return on investment (ROI). Pay close attention to the platform’s ability to improve threat detection accuracy and reduce false positives, as these factors directly impact SOC efficiency and effectiveness. Finally, look for support for emerging technologies such as generative AI and cloud security, which will help future-proof your security operations in an ever-evolving threat landscape.
The implementation of AI agents creates a streamlined, intelligent response pipeline that operates at machine speed, dramatically compressing traditional SOC timelines. By integrating AI agents into SOC operations, organizations benefit from a unified platform that consolidates security tools, data sources, and automation workflows, enabling more efficient management and enhanced visibility. This approach allows for rapid detection, contextual analysis, and response to any security incident, ensuring threats are addressed quickly and effectively.
AI agents receive alerts in real-time and immediately begin analysis. Using machine learning models trained on historical incident data, these agents can:
Once an alert is prioritized, AI agents simultaneously pull contextual information from multiple sources:
This parallel processing approach reduces enrichment time by over 90% compared to manual analyst workflows.
Based on enrichment findings, AI agents can either:
Every resolved incident feeds back into the AI agent's learning model, improving accuracy and reducing false positives over time. This creates a virtuous cycle where MTTR continues to improve as the system gains experience.
Consider a typical PowerShell-based malware investigation:
Traditional Approach (45 minutes):
AI Agent Approach (2 minutes):
Current industry benchmarks for SOC MTTR vary significantly by organization size and maturity:
Organizations implementing comprehensive AI agent solutions report dramatic MTTR reductions:
With routine tasks automated, security analysts can focus on:
Automation of repetitive tasks leads to:
Faster response times directly correlate with:
Successful AI agent deployment requires seamless integration with current security infrastructure:
SIEM Platform Integration: AI agents must consume and analyze data from Security Information and Event Management systems, requiring robust API connectivity and data normalization capabilities.
SOAR Platform Enhancement: Rather than replacing existing SOAR investments, AI agents should augment playbook execution with intelligent decision-making capabilities.
EDR/XDR Connectivity: Direct integration with Endpoint Detection and Response platforms enables real-time threat containment and detailed forensic analysis.
Organizations must determine their comfort level with AI agent autonomy:
Copilot Mode: AI agents provide recommendations and context, but human analysts make final decisions on remediation actions.
Semi-Autonomous Mode: AI agents can execute predefined "safe" actions (like blocking known-bad IPs) but require human approval for more significant interventions.
Fully Autonomous Mode: AI agents operate independently for routine threats, escalating only complex or high-risk scenarios to human analysts.
Simulation Environment: Begin with AI agents operating in read-only mode, comparing their recommendations against actual analyst decisions to build confidence in the system's accuracy.
Phased Rollout: Start with low-risk use cases (like automated alert enrichment) before progressing to containment actions and incident response.
Transparent Decision-Making: Ensure AI agents provide clear explanations for their actions, enabling human analysts to understand and validate the reasoning behind automated decisions.
VPC Deployment: Many organizations prefer deploying AI agents within their Virtual Private Cloud to maintain complete data control and comply with regulatory requirements.
On-Premises Options: For highly sensitive environments, on-premises AI agent deployment ensures that no security data leaves the organization's controlled infrastructure.
Hybrid Architectures: Combining cloud-based threat intelligence with on-premises processing can balance performance with privacy requirements.
A multinational organization faced escalating cybersecurity challenges:
Working with Arambh Labs, the organization deployed a comprehensive AI agent solution:
Phase 1 - Alert Triage (Month 1-2):
Phase 2 - Automated Enrichment (Month 3-4):
Phase 3 - Autonomous Response (Month 5-6):
MTTR Reduction:
Operational Efficiency:
Business Impact:
Key Success Factors:
The evolution of AI agents in cybersecurity is moving beyond reactive incident response toward proactive threat hunting and prevention. Next-generation AI agents will:
Future AI agents will function as tireless threat hunters, capable of:
Organizations with superior MTTR capabilities will gain significant competitive advantages:
As AI agent technology continues to mature, we can expect:
The cybersecurity landscape continues to evolve at breakneck speed, with threat actors becoming increasingly sophisticated and persistent. Traditional SOC approaches, while foundational, are no longer sufficient to defend against modern cyber threats. The organizations that will thrive in this environment are those that embrace intelligent automation and AI-driven security operations.
Arambh Labs is at the forefront of this transformation, helping organizations worldwide revolutionize their security operations through advanced agentic AI solutions. Our platform doesn't just reduce MTTR—it fundamentally transforms how security teams operate, making them more effective, efficient, and proactive in their defense strategies.
Don't let extended response times put your organization at risk. Discover how Arambh Labs' advanced agentic AI can transform your alert triage, enrichment, and remediation processes.
Contact us today to schedule a personalized demonstration and learn how AI agents can revolutionize your security operations. Your future self—and your security posture—will thank you.
8.9.2025 20:44Reduce SOC MTTR with AI AgentsThe 7 use cases for agentic AI in security operations include endpoint alert investigation, network alert investigation, identity-related investigation, cloud security incident handling, risk-based alert triage, advanced threat hunting, and insider threat detection & investigation. These applications help security teams achieve faster response times and reduce analyst burnout while transforming traditional security operations from reactive to proactive defense.
This comprehensive guide covers all 7 use cases with implementation benefits, real-world examples, and practical guidance for security teams. You’ll discover how agentic AI solutions address critical SOC challenges like alert fatigue, analyst shortages, and the need for 24/7 monitoring in modern IT environments.
Unlike traditional AI systems that require constant human intervention, these autonomous systems capable of independent decision making represent the next frontier in security operations, offering measurable business value through enhanced operational efficiency.
Agentic AI refers to autonomous systems capable of perceiving complex security situations, making intelligent decisions, and executing tasks independently without constant human oversight. These AI agents differ fundamentally from traditional security systems that follow predetermined rules or generative AI tools like ChatGPT that merely generate responses.
Key terminology includes:
Agentic AI operates autonomously by shifting security operations from reactive threat response to proactive defense. This transformation connects directly to SOC modernization through enhanced threat intelligence integration and streamlined incident response workflows.
The relationship works as follows: agentic AI systems → autonomous investigation → faster threat containment → reduced mean time to containment (MTTC). These AI capabilities enable continuous monitoring across diverse systems while adapting dynamically to emerging threats through advanced data analysis and natural language processing.
Current security teams face unprecedented challenges that make agentic AI solutions essential. According to recent market analysis, 40% of security leaders identify artificial intelligence systems as the biggest SOC impact driver over the next 12-24 months.
The numbers paint a clear picture of organizational pressure:
Agentic AI systems deliver quantifiable improvements. Companies implementing these solutions report significant reductions in mean time to containment (MTTC), transforming how security teams identify threats and respond to cyber threats. This represents a fundamental shift from manual processes toward automated incident response across enterprise systems.
Metric | Traditional SOC | Agentic AI-Powered SOC | Improvement |
|---|---|---|---|
Mean Time to Detection (MTTD) | 4-6 hours | 15-30 minutes | 85% reduction |
Mean Time to Containment (MTTC) | 2-4 days | 2-4 hours | 90% reduction |
False Positive Rate | 30-40% | 5-10% | 75% reduction |
Analyst Productivity | 20-30 alerts/day | 100+ alerts/day | 300% increase |
24/7 Coverage | Limited by staffing | Continuous monitoring | 100% uptime |
Threat Pattern Recognition | Rule-based only | Adaptive learning | Dynamic improvement |
AI agents enrich suspicious process or service creation, DLL injection, or persistence alerts with critical context such as hash lookups, parent-child process lineage, and MITRE ATT&CK mapping. This contextual enrichment enables faster and more accurate investigations of endpoint alerts, helping security teams quickly identify and contain threats.
Agentic AI systems can identify threats in real time and autonomously mitigate them by analyzing network traffic and user behavior. Agentic AI investigates anomalous network traffic, lateral movement patterns, or potential command-and-control (C2) callbacks by automatically building graph relationships between hosts, ports, and geolocation data. This comprehensive analysis uncovers sophisticated attack patterns that traditional tools may miss.
Agentic AI detects suspicious activities like privilege escalation, multi-factor authentication (MFA) bypass, impossible travel, or credential stuffing by cross-checking identity and access management (IAM) logs, login velocity, and user risk scores. This enables proactive identification of identity-based threats.
Agentic AI responds to cloud security incidents such as misconfigurations, unusual API calls, and privilege escalations in cloud platforms like AWS, GCP, and Azure. It analyzes CloudTrail, Stackdriver logs, policy changes, and IAM activity to detect and remediate cloud risks effectively. Agentic AI can continuously monitor and correct cloud misconfigurations and identity-based security issues to reduce attack surfaces.
AI agents prioritize raw security alerts by business risk, linking endpoint, identity, and cloud signals into a single ranked incident with an impact score. This risk-based triage reduces alert fatigue and ensures security teams focus on the most critical threats.
Agentic AI autonomously hunts across diverse datasets—including endpoint detection and response (EDR), network detection and response (NDR), IAM, and cloud logs—to identify stealthy techniques such as living-off-the-land binaries, beaconing patterns, or data staging activities. This proactive threat hunting uncovers hidden adversaries.
Agentic AI links human resources data, access logs, and file movement patterns to detect potential malicious insiders. It identifies behaviors like sudden sensitive data downloads, unusual login hours, or USB exfiltration, enabling early detection and mitigation of insider threats.
Mistake 1: Deploying Without Proper Governance Organizations often implement agentic AI solutions without establishing governance frameworks and maintaining appropriate human oversight. This can lead to unintended consequences or actions that disrupt business operations.
Mistake 2: Insufficient Integration Planning Failing to properly integrate agentic AI with existing security infrastructure and enterprise systems limits effectiveness and can create operational silos.
Pro Tip: Start with a single use case like alert triage to build confidence in AI capabilities, establish trust through transparency in decision making, then gradually expand to additional use cases as the system proves its value.
Case Study 1: Digital Insurance Company A major digital insurance provider implemented agentic AI for alert triage across AWS, Google Workspace, and Okta environments. The solution automatically correlates security alerts from multiple systems, reducing analyst workload by 85% while improving threat detection accuracy.
Results achieved:
Q1: Which use case should organizations implement first? Start with alert triage and investigation as it provides immediate relief from alert fatigue while building confidence in AI capabilities. This use case delivers quick wins and establishes the foundation for expanding to more complex autonomous systems.
Q2: How does agentic AI handle false positives in threat detection?
Advanced correlation algorithms and continuous learning reduce false positives by 70-80% compared to traditional rule-based systems. The AI agents learn from feedback loops and adapt their detection criteria based on confirmed threats versus benign activities.
Q3: What’s the typical ROI timeline for these use cases? Most organizations see measurable improvements within 3-6 months of successful implementation, with full ROI typically achieved within 12 months. The timeline depends on complexity of integration and organizational readiness.
Q4: Can agentic AI replace human security analysts entirely? No, agentic AI augments human capabilities by handling routine tasks and complex data analysis, allowing analysts to focus on strategic work, threat hunting, and complex decision making that requires human insight and creativity.
The 7 use cases for agentic AI in security operations represent a transformative shift from traditional security approaches to autonomous, intelligent defense systems. Organizations implementing these solutions achieve significant MTTC reduction and analyst productivity gains while addressing critical challenges like alert fatigue and analyst shortages.
Success requires starting with proper governance frameworks and gradual implementation. Begin with alert triage as your foundation use case, establishing trust and demonstrating value before expanding to automated incident response and autonomous threat hunting.
The benefits of agentic AI extend beyond operational efficiency to measurable business value through reduced cyber threat exposure, improved regulatory compliance, and enhanced security team effectiveness. As these autonomous systems continue evolving, organizations that embrace these use cases now will gain competitive advantages in threat detection and response capabilities.
Assess your current SOC maturity and identify which use case addresses your most pressing challenges. Whether facing alert fatigue, analyst shortages, or the need for 24/7 monitoring, agentic AI solutions offer proven approaches to transform your security operations for the modern threat landscape.
5.9.2025 19:207 Use Cases for Agentic AI in Security OperationsEarly Keyword Confirmation:
Evaluating agentic AI platforms for security operations is a critical process that helps SOC teams select autonomous AI solutions that can reduce Mean Time to Contain (MTTC) by up to 90% while enabling 24/7 threat detection and response. Unlike traditional security tools, these ai systems make independent decisions, analyze complex investigations, and execute tasks with minimal human input.
This comprehensive guide covers evaluation frameworks, key assessment criteria, vendor comparison methods, and implementation considerations for security operations centers looking to transform security operations through autonomous ai agents. You’ll learn how to assess platforms like Arambh Labs while avoiding costly implementation failures.
The evaluation process addresses immediate search intent for security leaders choosing between agentic ai systems that can handle overwhelming alert volumes, reduce alert fatigue, and enable faster threat detection without constant human oversight.
Agentic ai refers to artificial intelligence systems that possess autonomy, enabling them to independently analyze security alerts, reason through complex investigations, and take containment actions without human intervention. These platforms differ fundamentally from traditional ai and existing security tools by featuring multi agent ai systems where specialized agents handle different aspects of security operations.
Key terminology for evaluation includes:
Agentic ai systems integrate with existing security infrastructure through a layered approach: data ingestion from security tools → ai agent analysis using large language models → autonomous response execution → human oversight for validation. This represents a fundamental shift from traditional automation that requires human direction for each decision.
The integration map shows how agentic ai in security connects with:
Modern agentic ai agents must seamlessly integrate with existing security tools while maintaining the ability to execute tasks independently across hybrid cloud environments.
The impact of choosing the wrong agentic ai platform extends beyond implementation costs to operational effectiveness and security posture. Poor platform selection leads to delayed threat detection, continued false positives overwhelming soc analysts, and persistent analyst feedback about tool ineffectiveness that contributes to turnover.
Industry data reveals that 40% of security leaders expect artificial intelligence to significantly impact security operations centers within 12-24 months, making proper evaluation increasingly urgent. The cost implications are substantial - inadequate evaluation processes result in implementation failures averaging $2.4M in cybersecurity incidents due to gaps in threat detection and response capabilities.
Strategic importance centers on SOC transformation toward autonomous systems that can handle increasingly sophisticated threats without requiring additional human expertise. Organizations that properly evaluate agentic ai platforms position themselves to:
The benefits of agentic ai become measurable only when platforms are properly evaluated against specific organizational needs and integrated thoughtfully with existing security infrastructure.
Evaluation Factor | Traditional AI Tools | Agentic AI Systems | Measurement Criteria |
|---|---|---|---|
Autonomy Level | Rule-based responses | Independent decision-making | % investigations completed without human intervention |
MTTC Reduction | 20-40% improvement | 80-90% improvement | Time from alert to containment |
Integration Depth | API connections only | Native multi-agent coordination | Compatibility with existing tools |
Transparency | Limited audit trails | Complete investigation documentation | Explainability of AI decisions |
False Positive Handling | Static filtering rules | Adaptive learning from analyst feedback | Reduction in low priority alerts |
Performance Benchmarks for Evaluation:
Cost-Benefit Analysis Framework: Calculate ROI using current analyst costs ($150K+ annually), time spent on routine tasks (60-80% of analyst hours), and cost of security incidents (average $4.45M per data breach). Factor in reduced need for 24/7 staffing and improved analyst retention through eliminating repetitive tasks.
Technical Requirements Matrix:
Begin evaluation by auditing your existing security tool stack, daily alert volumes, and how soc analysts currently distribute their time across routine tasks versus complex investigations. Document current MTTC, false positive rates, and analyst feedback about alert fatigue to establish baseline metrics.
Define Success Criteria:
Create Requirement Checklist:
Test autonomous decision-making capabilities through proof-of-concept scenarios using your actual security data, not sanitized vendor demonstrations. Focus on how ai agents handle edge cases, coordinate investigations across multiple data sources, and provide transparent reasoning for their conclusions.
Assess Agent Specialization:
Evaluate Transparency Features:
Recommended Evaluation Tools:
Establish comprehensive baseline metrics covering current MTTC, false positive escalation rates, and analyst efficiency scores before platform deployment. Run comparative tests measuring threat detection accuracy, investigation thoroughness, and response speed using controlled scenarios that mirror your typical threat landscape.
Performance Validation Process:
Deployment Timeline Benchmarks:
Track roi metrics including reduced analyst overtime, improved retention rates, and faster containment of sophisticated threats that previously required extensive human expertise.
Mistake 1: Focusing solely on AI capabilities without thoroughly assessing integration complexity with existing security infrastructure. Many organizations underestimate the effort required to connect agentic ai systems with legacy SIEM platforms, custom security tools, and hybrid cloud environments.
Mistake 2: Ignoring transparency and auditability requirements that are critical for regulatory compliance and building analyst trust. Security teams must be able to understand and validate ai agent decisions, especially for high-impact containment actions affecting customer data or business operations.
Mistake 3: Underestimating change management needs when transitioning from manual investigation processes to autonomous systems. Analyst resistance often stems from fear of job displacement rather than understanding how agentic ai enhances human expertise rather than replacing it.
Pro Tip: Always evaluate platforms using your actual security data and realistic threat scenarios, not polished vendor demonstrations. Request access to sandbox environments where you can test edge cases, integration challenges, and decision transparency using your organization’s specific threat landscape and compliance requirements.
Additional evaluation pitfalls include rushing deployment timelines, failing to establish clear success metrics, and not planning for ongoing analyst feedback mechanisms that enable continuous improvement of ai agent performance.
Case Study: Fortune 500 financial services company reduced MTTC from 4 hours to 5 minutes using systematic agentic AI platform evaluation methodology.
Starting Situation:
Evaluation Process - 6-Week Assessment:
Week 1-2: Requirements gathering and baseline establishment
Week 3-4: Platform testing with three vendors
Week 5-6: Proof-of-concept validation using actual incident data
Final Results After 12-Week Implementation:
Metric | Before Agentic AI | After Implementation | Improvement |
|---|---|---|---|
MTTC | 4 hours | 5 minutes | 90% reduction |
False Positives | 85% | 20% | 75% elimination |
Analyst Efficiency | 25% complex work | 80% complex work | 60% improvement |
Alert Fatigue Score | 8.5/10 | 3.2/10 | 65% reduction |
The selected platform Arambh Labs demonstrated superior autonomous investigation capabilities while maintaining complete audit trails for regulatory compliance. Integration with existing security infrastructure required minimal disruption, and analyst feedback showed high satisfaction with reduced repetitive tasks and increased focus on strategic threat hunting activities.
Q1: How long does a typical agentic AI platform evaluation take?
A1: Most comprehensive evaluations require 8-12 weeks including proof-of-concept testing, vendor comparisons, and pilot deployment validation. This timeline allows for thorough testing of autonomous ai systems with your actual security data and integration requirements.
Q2: What’s the difference between evaluating agentic AI vs traditional SIEM platforms?
A2: Agentic ai evaluation focuses on autonomous decision-making capabilities, multi-agent coordination, and investigation transparency rather than rule configuration and workflow optimization. You’re assessing artificial intelligence systems that can execute tasks independently, not tools requiring constant human direction.
Q3: Should we evaluate cloud-native or on-premises agentic AI platforms?
A3: Cloud-native platforms typically offer faster deployment (4-12 weeks vs 6+ months) and self-tuning capabilities that reduce ongoing maintenance. They also provide better scalability for handling overwhelming alert volumes and integration with modern cloud environments where most security tools now operate.
Q4: How do we measure the benefits of agentic ai during evaluation?
A4: Focus on quantifiable metrics like MTTC reduction, false positive elimination, and analyst time allocation changes. Track how well ai agents handle sophisticated threats that previously required extensive human expertise, and measure improvements in analyst feedback about job satisfaction.
Q5: What integration challenges should we expect during evaluation?
A5: Common challenges include API compatibility with legacy security tools, data protection requirements for customer data, and ensuring behavioral analytics work effectively with your specific network architecture. Plan for 4-8 weeks of optimization to fine-tune ai agent performance for your environment.
Successful agentic AI platform evaluation requires prioritizing true autonomy with specialized agents for tier 1 and tier 2 operations rather than AI-assisted tools requiring constant human oversight. Focus on platforms that demonstrate independent decision-making, multi-agent coordination, and the ability to handle increasingly sophisticated threats without overwhelming your soc team with false positives.
Critical Success Factors:
Next Action Steps:
Download vendor evaluation checklists that include technical requirements for your security tools, schedule proof-of-concept demonstrations with shortlisted platforms using your actual alert data, and begin baseline metric collection to measure roi from autonomous systems implementation. Remember that the goal is transforming security operations to match the pace and sophistication of modern threats while reducing alert fatigue and improving analyst satisfaction.
The shift toward agentic ai in security represents a fundamental evolution from reactive to proactive security operations, enabling security teams to focus human expertise on strategic initiatives while ai agents handle the overwhelming volume of routine security alerts and initial investigations.
3.9.2025 21:51How to Evaluate Agentic AI Platform for Security OperationsAI will not replace cybersecurity professionals but will transform their roles and enhance their capabilities in the evolving threat landscape. While artificial intelligence is revolutionizing how we approach cyber security, the human element remains irreplaceable for strategic thinking, ethical decision-making, and complex threat analysis.
This comprehensive guide covers AI’s current role in cybersecurity, job transformation versus replacement, emerging opportunities, and preparation strategies for the 3.5 million cybersecurity professionals worldwide who are concerned about AI’s impact on their careers.
The timing of this discussion is critical. With cyber attacks occurring every 39 seconds and AI adoption accelerating across security teams, understanding how to work alongside AI systems rather than compete with them has become essential for cybersecurity careers. The cybersecurity industry faces a massive skills gap with millions of unfilled positions, creating unprecedented opportunities for professionals who embrace AI as a collaborative tool.
Artificial intelligence in cybersecurity refers to LLMs, machine learning algorithms, natural language processing, and automated systems that enhance threat detection, incident response, and vulnerability management. These AI technologies work by analyzing massive datasets to identify patterns, anomalies, and potential security threats that might overwhelm human analysts.
Key terminology includes:
AI excels at processing raw data at scale, and now with LLMs reasoming, it has better understanding novel attack vectors that require human expertise and creative problem-solving.
Modern cybersecurity systems integrate AI across multiple domains:
Real-time threat detection uses machine learning to analyze network traffic, identifying suspicious patterns that indicate potential breaches. These ai systems can process millions of events per second, flagging anomalies for human investigation.
Automated vulnerability scanning employs AI to continuously assess computer systems for security weaknesses, prioritizing patches based on threat intelligence and business impact. This reduces analysts workload while maintaining comprehensive security coverage.
AI-powered tools like Arambh Labs' agentic platform, Microsoft Security Copilot and CrowdStrike Falcon have transformed security operations centers. These platforms combine machine learning with human oversight to accelerate incident management and reduce false positives by up to 70%.
Predictive analytics for zero-day attacks uses generative ai to model potential attack scenarios based on current threat landscapes, helping security teams prepare defenses for future threats before they materialize.
Statistical evidence demonstrates that AI-human collaboration produces superior results compared to either approach alone. Organizations using ai driven security solutions report 60% faster threat detection and resolution times while maintaining human oversight for critical decisions.
The cybersecurity skills gap creates a compelling case for AI augmentation rather than replacement. With 3.5 million unfilled cybersecurity jobs projected by 2025, ai tools enable existing security teams to cover more ground effectively. Rather than eliminating positions, AI is helping organizations scale their cybersecurity efforts without proportional headcount increases.
Cost-benefit analysis from major financial institutions shows that AI enhances productivity by automating routine tasks like log analysis and initial threat triage, freeing cybersecurity professionals to focus on strategic initiatives, complex investigations, and stakeholder communication that require human judgment.
Real-world examples from enterprise security implementations demonstrate successful AI-human collaboration. Organizations report that while AI handles high-volume data analysis and pattern recognition, human experts provide strategic direction that ai models cannot replicate.
Current ai systems struggle with sophisticated attacks that use social engineering, zero-day exploits, or adversarial techniques designed to fool machine learning algorithms. These scenarios require human expertise to analyze attacker motivations, predict next moves, and develop countermeasures.
The false positive challenge remains significant, with even advanced ai models generating alerts that require human validation. Organizations report that 30-40% of AI-flagged incidents need human analysis to determine legitimacy, highlighting the continued need for skilled analysts.
Ethical and privacy concerns in AI decision-making cannot be automated. When ai systems recommend blocking certain network traffic or quarantining user accounts, human oversight ensures that these actions align with business objectives and regulatory requirements.
Entry-level SOC analyst positions are evolving rather than disappearing. Instead of manually monitoring security alerts, modern analysts work with ai powered tools to investigate complex incidents, validate automated responses, and develop new detection rules based on emerging threat patterns.
New responsibilities for cybersecurity professionals include:
Skills that remain uniquely human include creative problem-solving when facing novel attack vectors, ethical judgment in balancing security with business operations, and stakeholder communication to translate technical risks into business language.
AI Security Analysts combine traditional cybersecurity knowledge with machine learning expertise. These professionals earn 15-25% more than traditional analysts, with responsibilities including ai model oversight, bias detection, and strategic threat analysis. Required qualifications include cybersecurity fundamentals plus training in data analysis and machine learning principles.
Machine Learning Security Engineers focus on protecting ai systems themselves from adversarial attacks while implementing AI-driven security solutions. This role requires deep technical knowledge of both cybersecurity and AI technologies, with career progression toward senior architect positions.
AI Ethics and Compliance Specialists ensure that ai driven security solutions meet regulatory requirements and organizational values. These professionals navigate the complex intersection of AI capabilities, privacy regulations, and business ethics.
Cybersecurity AI Trainers develop and maintain training data for security ai models, ensuring that algorithms learn from diverse, representative datasets while avoiding bias that could create security blind spots.
Technical skills for the AI era include:
Strategic skills encompass:
Communication skills become increasingly important as professionals must explain AI decisions to non-technical stakeholders, justify AI-driven security investments, and coordinate between technical teams and business leadership.
Misconception 1: AI can handle all threat detection autonomously without human validation. Reality: Even the most advanced ai systems require human oversight for complex incidents, false positive management, and strategic decision-making. Organizations using fully automated systems report higher security risks due to lack of contextual understanding.
Misconception 2: AI eliminates the need for cybersecurity expertise and strategic thinking. Reality: AI tools require skilled operators who understand both cybersecurity principles and AI capabilities. The technology amplifies human expertise rather than replacing it.
Misconception 3: AI adoption will lead to mass unemployment in cybersecurity. Reality: The cybersecurity skills gap means that ai tools help organizations do more with existing staff rather than reducing headcount. Most organizations report redeploying staff to higher-value activities rather than eliminating positions.
Pro Tip: Reframe AI concerns as career advancement opportunities. Professionals who develop AI skills early position themselves for leadership roles in an increasingly AI-augmented cybersecurity landscape. The job market rewards those who can bridge traditional cybersecurity knowledge with modern AI capabilities.
Q: Will entry-level cybersecurity jobs disappear due to AI? A: Entry-level roles will evolve to include AI collaboration skills, creating new opportunities for tech-savvy professionals. Organizations need junior staff who can work with ai systems while developing traditional cybersecurity expertise.
Q: How long do I have to adapt to AI in cybersecurity? A: The transformation is gradual; professionals have 2-3 years to develop AI-related skills while demand grows. Early adopters gain competitive advantages in the job market.
Q: What certifications should I pursue for AI-enhanced cybersecurity? A: Consider Google Cybersecurity Certificate with AI components, IBM Generative AI for Cybersecurity, and vendor-specific training for ai powered tools like Microsoft Security Copilot and CrowdStrike Falcon.
Q: Can AI handle advanced persistent threats without human input? A: No, APTs require human analysis for context, attribution, and strategic response planning. AI provides valuable data analysis, but human expertise remains essential for understanding attacker motivations and developing comprehensive defense strategies.
Q: Should I be concerned about AI creating new cybersecurity risks? A: Yes, but these risks create opportunities for security professionals. Attackers ai capabilities require defenders who understand both traditional threats and AI-specific vulnerabilities, expanding rather than contracting career opportunities.
Five key takeaways shape the future of cybersecurity careers: AI augments rather than replaces human expertise, new specialized roles are emerging rapidly, human skills like creativity and ethical reasoning remain critical, continuous learning becomes essential for career growth, and opportunities significantly outweigh threats for adaptable professionals.
Take immediate action by starting to learn ai tools relevant to your cybersecurity role, pursuing AI-focused certifications, and embracing artificial intelligence as a career accelerator rather than a threat. The professionals who thrive will be those who master the collaboration between human expertise and AI capabilities.
The cybersecurity industry will continue growing as ai systems create more sophisticated attack vectors and defense requirements. Rather than eliminating cybersecurity jobs, AI is creating more rewarding career paths that combine traditional security knowledge with cutting-edge technology skills.
Begin your AI-cybersecurity education journey today to stay ahead of industry evolution. The future belongs to security professionals who can leverage ai technologies while providing the strategic thinking, ethical guidance, and creative problem-solving that only humans can deliver.
26.8.2025 21:00Will AI Replace Cyber Security Professionals? The Future of Cybersecurity CareersSecurity operations automation is rapidly evolving with the integration of Large Language Models (LLMs), transforming how organizations detect, respond to, and mitigate security threats. LLMs bring advanced natural language understanding and contextual analysis capabilities to security automation platforms, enabling more accurate threat detection, enhanced incident response, and reduced alert fatigue for security teams.
This guide explores how LLM-powered security operations automation redefines cybersecurity workflows, streamlines complex security tasks, and empowers security analysts to focus on strategic initiatives by minimizing manual intervention in routine and repetitive security tasks.
Large Language Models are advanced artificial intelligence systems trained on vast amounts of textual data to understand, generate, and analyze human language with high accuracy. In cybersecurity, LLMs can process and interpret complex security data, logs, and alerts, providing deep insights and automating tasks that traditionally required extensive human expertise.
LLMs augment traditional security automation tools by enabling:
LLM-driven automation platforms seamlessly integrate with existing security tools such as SIEM, SOAR, and XDR, orchestrating workflows across multiple security layers to provide a unified and intelligent security posture.
Modern security teams face an overwhelming volume of security alerts and increasingly sophisticated cyber threats. The cybersecurity skills shortage exacerbates these challenges, making it difficult for analysts to keep pace with manual security processes.
LLM-powered automation addresses these issues by:
Identify manual security processes and routine tasks that can benefit from LLM-powered automation, such as alert triage, threat intelligence enrichment, and incident response workflows.
Choose platforms that offer seamless integration with your existing security tools and support dynamic learning and adaptation to evolving threats.
Leverage LLM capabilities to build intelligent playbooks that automate complex investigations, enrich alerts with contextual data, and execute precise response actions with minimal human intervention.
Continuously assess automation effectiveness through key performance indicators like mean time to detection (MTTD), mean time to response (MTTR), false positive rates, and analyst productivity. Refine workflows and expand automation scope to include advanced threat hunting and predictive analytics.
LLM agents analyze email content and metadata to identify phishing attempts, automatically quarantining suspicious messages and notifying security teams with detailed context.
LLM agents aggregate data from multiple sources, providing enriched threat intelligence that helps prioritize alerts and guides incident response.
LLM agents autonomously investigate security alerts, correlate related events, and trigger automated response playbooks, reducing response times and improving accuracy.
While LLMs offer transformative benefits, organizations must address challenges such as data privacy, model explainability, and integration complexity. Best practices include maintaining human oversight for critical decisions, ensuring continuous model training with up-to-date threat data, and prioritizing seamless integration with existing security infrastructure.
Integrating Large Language Models into security operations automation marks a paradigm shift in cybersecurity. By combining the power of LLMs with existing security automation tools, organizations can achieve faster, more accurate threat detection and response while mitigating alert fatigue and operational inefficiencies.
Security operations automation with LLMs empowers security teams to stay ahead of emerging threats, enhancing overall security posture and resilience in an increasingly complex cyber threat landscape. Embracing this technology today is essential for organizations aiming to maintain robust, scalable, and intelligent security operations.: Complete Guide to Streamlining Cybersecurity in 2024
Q1: What’s the difference between LLM based security automation and traditional security automation platforms aka SOAR? A1: LLM-based security automation goes beyond pre-scripted playbooks by understanding natural language, reasoning over complex alerts, and dynamically generating investigation or remediation steps. Traditional SOAR platforms rely on static, rule-based workflows that must be manually defined and updated for every new threat pattern. In contrast, LLM-driven systems adapt in real time, reduce engineering overhead, and handle novel or unstructured security data that SOARs typically miss.
Q2: How does AI enhance security operations automation beyond rule-based systems? A2: Artificial intelligence enables dynamic decision-making, learns from historical security incidents, and adapts to new threat patterns without requiring manual updates to detection rules and response playbooks.
Q3: Can small security teams benefit from automation or is it only for large enterprises? A3: Small security teams often see the greatest ROI from security automation as it helps them scale operations and improve security posture without proportional increases in headcount or operational costs.
Q4: How do automated systems handle emerging threats that haven’t been seen before? A4: Modern security automation platforms use machine learning algorithms and behavioral analysis to detect anomalous patterns that may indicate new attack methods, complementing signature-based detection with adaptive threat detection capabilities.
Q5: What level of human intervention should remain in automated security processes? A5: Critical security decisions, complex incident investigations, and situations requiring business context should maintain human oversight, while routine security tasks like alert enrichment and standard response actions can be fully automated.
21.8.2025 21:30Security Operations Automation with LLMs: Revolutionizing Cybersecurity in 2025