The AI Automation Arms Race: Why Defense Is Not Symmetrical

The security industry likes to tell itself a comforting story: as attackers adopt artificial intelligence, defenders will respond in kind, and the balance of power will remain roughly equal. AI on both sides, the thinking goes, should cancel out.

This assumption is wrong — and potentially dangerous.

As with most other areas of security, the use of AI in practice is deeply asymmetrical. Attackers benefit disproportionately from automation, while defenders struggle to translate AI adoption into meaningful risk reduction. The result is not an arms race between equals, but a widening gap between the speed at which attacks evolve and the pace at which enterprises can govern, understand, and respond to them.

Automation Favors the Offense

Attackers have always benefited from scale and, unfortunately, AI simply amplifies that advantage.

Modern attack campaigns will utilize automation not to invent entirely new exploits, but to industrialize existing ones. Known vulnerabilities, misconfigurations, and weak identity controls are now stitched together by AI-assisted tooling that adapts quickly, probes relentlessly, and exploits opportunity at machine speed. These campaigns are quieter, more persistent, and harder to distinguish from background noise.

Critically, attackers do not need perfect precision. They benefit from volume, iteration, and probabilistic success. A small improvement in targeting or evasion, multiplied across thousands of attempts, yields meaningful results and AI excels at this kind of iterative optimization.

Why Defensive AI Struggles to Keep Up

On the defensive side, AI is often deployed as an enhancement to existing tools, offering faster detection, better prioritization, and smarter correlation within an enterprise SIEM, for example. These are valuable improvements, but they do not change the fundamental constraints under which defenders must operate.

Security teams are accountable for outcomes, needing to explain decisions, justify actions, and demonstrate control. False positives disrupt business. False negatives create risk. Every automated response must be defensible after the fact to management, auditors, regulators, or customers.

This asymmetry matters. An attacker can afford to be wrong repeatedly, trusting that eventually a shot will be on target. Defenders, just like the poor old goalkeeper in soccer, cannot afford to be wrong once.

AI-powered detection may surface more signals, but without clear governance, visibility, and control, those signals quickly become noise. Automation without accountability simply accelerates confusion.

The Persistence of “Old” Attack Surfaces

One of the more uncomfortable realities emerging from recent incidents is that many AI-enabled attacks still rely on very traditional weaknesses, such as exposed services, misconfigured cloud environments, weak access controls, unpatched software, and multiple evasion techniques. AI does not replace these attack vectors but instead makes them easier to discover and exploit at scale.

To any seasoned security professional this sounds very familiar. The old saying goes: “there is nothing new under the sun” and the danger is not that AI introduces entirely new classes of risk overnight (although it inevitably will – see our white papers AI Security Beyond the Model: What Enterprises Need to Care About — and Whyand Evaluating Enterprise AI Security: Questions Every Buyer Should Be Able to Answer”). The real danger in the short term is that it quietly magnifies existing weaknesses until they become systemic failures.

Why Symmetry Is the Wrong Mental Model

The idea of a balanced AI arms race assumes that both sides gain comparable benefits from automation. In reality, the incentives and constraints are fundamentally different.

Attackers optimize for opportunity and speed, while defenders optimize for stability, correctness, and trust. AI aligns naturally with the former, aligning with the latter only when paired with strong governance, observability, and control.

This is why simply “adding AI” to security tools does not meaningfully close the arms race gap. Without clear policies, auditability, and predictable behavior under stress, automation can undermine confidence rather than strengthen it.

Reframing the Defensive Objective

The goal of AI security should not be to match threat actors’ algorithm for algorithm, since that race is unwinnable and misses the point.

AI-enabled systems must be designed and evaluated not just for detection capability, but for how they behave when assumptions break, when inputs are ambiguous, dependencies fail, or automation makes the wrong decision.

This requires shifting emphasis away from novelty and toward discipline and focusing on visibility into how decisions are made, evidence that controls behave predictably under stress, and governance that aligns automation with enterprise risk tolerance.

What CISOs Should Do Now

  • Treat AI-enabled security controls as governed systems, not intelligent features. Demand clarity on how automation is authorized, constrained, and audited.
  • Insist on observability and accountability. If your team cannot reconstruct why an automated decision was made, you cannot defend it to regulators, boards, or customers.
  • Pressure-test failure modes. Ask how controls behave when dependencies degrade, data is ambiguous, or models behave unexpectedly.
  • Resist the temptation to equate speed with strength. Automation that outpaces governance increases risk rather than reducing it.

Enterprises that succeed in this environment will not be those that deploy the most AI, but those that govern it best.

The uncomfortable truth is that AI makes security failures more likely when organizations do not understand their own systems. The answer is not less automation, but better control over how automation is introduced, evaluated, and held accountable.

The arms race metaphor obscures this reality. Defense is not about symmetry, but responsibility and accountability. And in an AI-enabled world, responsibility is the hardest capability to automate.

When AI Finds Every Bug

The discovery clock just accelerated

On April 7, Anthropic announced that its newest model, Claude Mythos Preview, had autonomously discovered thousands of high and critical severity zero-day vulnerabilities across every major operating system and web browser—many hiding in plain sight for over a decade. A 27-year-old bug in OpenBSD. A 16-year-old flaw in FFmpeg that automated fuzzers had hit five million times without catching. And Mythos doesn’t just find vulnerabilities—it writes the exploits, succeeding on over 83% of first attempts where previous models achieved close to zero.

That is good news in one sense: software vendors and maintainers may be able to identify and patch flaws earlier, and reduce the time that dangerous defects remain unknown.

But that same acceleration also accelerates threat landscape changes for everyone else. The same class of AI capability that can help defenders find weaknesses can also help bad actors understand them faster, build exploits around them faster, and operationalize attacks at far greater speed and scale.

This is a watershed moment. But while the headlines focus on offensive implications, the downstream consequences for enterprise security operations are just as profound.

The deployment clock has not

The optimistic reading is simple: vulnerabilities found earlier get patched earlier. Project Glasswing—Anthropic’s well thought-out consortium with AWS, Apple, Cisco, Google, Microsoft, and others—is already putting Mythos to work scanning critical codebases. But enterprise security leaders know the uncomfortable truth: a patch being available and a patch being deployed are two very different things.

Even when software owners produce fixes more quickly, enterprise deployment lifecycles still have to contend with regression testing, change windows, operational dependencies, rollback planning, uptime requirements, and the broader risks that come with touching mission-critical systems.

For most enterprises, especially those operating essential services or critical infrastructure, upgrade patterns will still need to follow established risk-mitigation best practices. The cost of an unstable production change can be just as severe as the vulnerability itself.

Layered defenses under increasing pressure

That leaves a familiar but increasingly compressed exposure gap: the period between when a vulnerability is known and when an organization can safely deploy the patched version into production.

As has been true for years, enterprises will need to rely on layered defenses to help close that gap. Firewalls, IDS/IPS, segmentation controls, and related security systems will remain essential in providing mitigation protection while upgrades are planned, tested, and rolled out safely.

What changes now is the pace and volume. If AI sharply increases the rate at which vulnerabilities are discovered, then the burden on protective controls will grow with it. Those systems will need to implement and deploy signatures, policies, and other mitigation mechanisms more quickly and more often.

Deployment has to be not just faster, but smarter

In this environment, simply deploying a mitigation faster is no longer enough. A signature added to an IPS, a rule pushed to a firewall, or a policy configured on paper does not by itself prove that the control is effective against the exploit path it is supposed to stop. An overly broad signature triggering false positives will interrupt legitimate business.

Those mitigation measures can be made smarter through validation that they are effective in protecting against the consequential vulnerabilities and exploits being targeted. Otherwise, organizations are not demonstrating risk reduction; they are assuming it.

This distinction will matter more as AI also helps adversaries accelerate exploit development and surrounding tooling. The attack side of the equation will move faster, which means control effectiveness must be measured more often and with greater rigor. Absent that, we will have a case where enterprises try to move fast but end up breaking things.

Why oversight pressure will rise

A hyper-accelerated threat environment will not only affect security teams. Lawmakers, regulators, insurers, auditors, boards, and other oversight bodies focused on enterprise risk are watching the same headlines. When AI can autonomously compromise critical infrastructure software, tolerance for vague assurances evaporates.

Their questions will increasingly shift from broad policy statements to operational evidence. Questions will evolve from “Do you have security controls?” to “What have you done to prevent this, and what is the evidence it’s working?”.

As a result, requirements for proof of continuous validation testing and evaluation of deployed security controls are likely to become more common across corporate governance, risk management, compliance, insurance underwriting, and sector-specific oversight agencies.

Continuous validation becomes a core operating discipline

This is why continuous validation testing of security controls will become more critical in enterprise IT operations.

Continuous validation gives organizations a practical way to bridge the gap between faster vulnerability discovery and the slower, smarter, necessary discipline of safe production change. It helps prove that compensating controls are providing real mitigation while the enterprise works through responsible remediation cycles.

In the age of AI-accelerated vulnerability discovery and exploit creation, continuous validation will become central not only to security operations, but also to meeting corporate GRC obligations, supporting regulatory readiness, and demonstrating defensible cyber resilience.

Organizations that build this into their operational cadence, GRC programs, and vendor accountability frameworks will be able to answer the hard questions when regulators and boards come asking. Those that don’t will find themselves defending assumptions in a world that no longer accepts them.

Continuous validation isn’t the future of enterprise security operations. It’s the present. Mythos just made it impossible to ignore, and Project Glasswing presents a great opportunity to respond.

What enterprise leaders should do now

Priority Why it matters
Preserve disciplined change management Do not trade mission-critical availability for superficial patch velocity. Upgrades still need controlled testing and rollout.
Strengthen layered defenses Firewall and IDS/IPS mitigations will be asked to carry more of the burden during the exposure window.
Validate mitigations continuously Controls need evidence-based testing to prove they actually block the vulnerabilities and exploits they target.
Prepare for evidence demands Regulators, insurers, boards, and auditors will increasingly expect proof, not just claims, that controls are working.

Inside Cybersecurity: NSS Labs Issues Whitepapers on Enterprise AI Governance, Launches Testing Initiative

Cybersecurity testing firm NSS Labs has published a series of whitepapers to help organizations address artificial intelligence security governance and has launched a new program to evaluate the effectiveness of AI protection systems.

“We’re at the beginning of the AI revolution and everyone has questions. These papers provide a framework for how to think about securing AI as well as practical guidance for governance of what their AI systems are permitted to do and why. Yes, AI security is a technical issue, but it is also a governance issue,” Vikram Phatak, CEO of NSS Labs, said in a March 18 press release.

The first whitepaper from NSS Labs lays out the argument for enterprise AI security to be “treated as a system-level and governance challenge,” according to the release.

The second whitepaper provides topic areas and questions buyers should ask to evaluate the security of an AI product.

Read the full article here.

NSS Labs Appoints Industry Veteran Dominick Delfino as Executive Advisor

Austin, TX – March 24, 2026 – NSS Labs, the leading authority in independent cybersecurity product validation, today announced the appointment of Dominick Delfino as Executive Advisor. A seasoned technology leader with more than 25 years of experience at Google Cloud, Nutanix, Pure Storage, and Cisco, Delfino will provide strategic guidance to the NSS Labs leadership team as the company expands its testing capabilities for the next generation of AI-driven cybersecurity.

Delfino joins NSS Labs at a pivotal moment for enterprises, where the rise of sophisticated, automated threats has made independent, real-world validation of security efficacy more critical than ever.

Most recently, Delfino served as Global Vice President of Cybersecurity Sales at Google Cloud, where he led the global go-to-market strategy for the company’s security portfolio, including the integration of Mandiant. His distinguished career also includes serving as Chief Revenue Officer at Nutanix and Pure Storage, as well as holding senior leadership roles at VMware and Cisco.

“Dominick is a distinguished leader in the technology and security space,” said Vikram Phatak, CEO of NSS Labs. “His experience scaling global organizations and his deep understanding of the cloud and security landscape from his time at Google Cloud and VMware will be invaluable. Dominick understands exactly what enterprise customers need, and his guidance will be instrumental as we grow our enterprise programs.”

“Throughout my career, I’ve witnessed how difficult it is for organizations to separate marketing claims from actual security performance,” said Delfino. “NSS Labs has always stood for transparency and data-driven truth in a crowded marketplace. I am thrilled to be helping the team scale and ensure that enterprises have the right tools to deliver independent, real-world validation of their security controls.”

As Executive Advisor, Delfino will focus on accelerating NSS Labs global sales, enhancing strategic partnerships, and aligning the company’s roadmap with the rapidly shifting requirements of AI.

NSS Labs Names Keysight Lead Partner in New AI Protection Systems Security Testing Initiative

Austin, TX — March 23, 2026. NSS Labs today announced that Keysight Technologies has joined its new AI Protection Systems (AIPS) security testing initiative as lead partner, supporting the development of one of the industry’s first independent evaluation programs dedicated to testing AI security guardrail technologies.

As artificial intelligence becomes foundational to digital transformation across industries—including finance, healthcare, government, and critical infrastructure—the security and integrity of AI systems has emerged as a global priority. Organizations are rapidly deploying AI models and applications, yet the technologies designed to secure and govern their use—often referred to as AI guardrails, AI firewalls, or AI runtime protection systems—have not yet been independently validated through standardized testing.

To address this gap, NSS Labs is launching a comprehensive independent evaluation program dedicated specifically to AI Protection Systems (AIPS)—security platforms designed to enforce policy, prevent misuse, and defend AI models and applications from adversarial attacks. The initiative aims to establish a transparent, technically rigorous methodology that benchmarks how effectively these systems protect AI deployments against real-world threats while maintaining policy enforcement and operational integrity.

Keysight is a foundational partner, supporting the development and execution of this groundbreaking AI security validation program.

The NSS Labs AIPS methodology evaluates products across seven dimensions of AI security, including malicious input and prompt attacks, output risks and sensitive data exposure, system resilience under adversarial conditions, policy enforcement accuracy, agentic AI and tool invocation security, observability and audit capabilities, and performance and scalability impact.

Across these layers, the methodology includes hundreds of thousands of individual test case executions designed to bypass, manipulate, exploit, or overwhelm AI Protection Systems. Each scenario is executed using multiple attack samples and variations—including prompt injection attempts, jailbreak techniques, obfuscated prompts, sensitive data extraction attempts, exploit generation requests, RAG poisoning attacks, API privilege escalation attempts, and agent tool misuse scenarios.

“AI is rapidly becoming core infrastructure for the digital economy, and with that comes an urgent need for independent validation of the technologies designed to protect it,” said Vikram Phatak, CEO of NSS Labs. “With Keysight joining us as lead partner, we are bringing together our deep expertise in testing along with Keysight’s global innovation solutions that will help the industry understand how well AI protection systems actually perform against real-world threats.”

“AI is quickly becoming foundational infrastructure, and trust in these systems must be earned through transparent, independent validation,” said Ram Periakaruppan, Vice President and General Manager, Network Test & Security at Keysight. “Keysight’s strength in building scalable, real-world test environments and generating actionable performance insights positions us to help shape how AI security is measured. We’re proud to partner with NSS Labs to advance a more resilient and trustworthy AI ecosystem.”

By combining adversarial testing, policy validation, system robustness analysis, and operational visibility checks into a single structured framework, the NSS Labs AI Protection Systems test aims to establish a credible benchmark for independent AI security validation and provide enterprises with objective data on the effectiveness of technologies designed to secure AI.

Feedback for the methodology is currently being accepted from enterprises and security vendors. Please reach out to [email protected] for a draft copy if you would like to provide comments. The AIPS methodology will be published in April.

Executives from NSS Labs are attending the RSA Conference. Please contact us if you would like to schedule a meeting.

NSS Labs Publishes Two Foundational White Papers on Enterprise AI Security

Austin, TX – March 18, 2026 – NSS Labs, the leading authority in independent cybersecurity product validation, today announced the publication of two new white papers addressing the rapidly evolving challenge of securing artificial intelligence in enterprise environments:

Together, the papers provide enterprise security leaders with a structured, governance-driven framework for understanding AI risk in production systems. The research was developed in collaboration with Amazon Web Services (AWS), F5, and Microsoft as well as other industry leaders.

AI Security Beyond the Model: What Enterprises Need to Care About — and Why,” outlines why securing the AI model alone is insufficient and why enterprise AI security must be treated as a system‑level and governance challenge. The aim is to provide concrete guidance to Chief Information Security Officers (CISOs), enterprise buyers, and Governance, Risk and Compliance (GRC) leaders on the questions to ask before real-world AI failures are exposed under regulatory, legal, customer, or board-level scrutiny.

“Evaluating Enterprise AI Security: Questions Every Buyer Should Be Able to Answer” moves from theory to procurement discipline to help enterprise buyers formulate better questions when shortlisting AI security vendors. The focus is primarily on runtime guardrails in the form of AI Protection Systems, the controls outside the model that enforce policy, protect data, and produce audit evidence.

“We’re at the beginning of the AI revolution and everyone has questions,” said Vikram Phatak, CEO of NSS Labs. “These papers provide a framework for how to think about securing AI as well as practical guidance for governance of what their AI systems are permitted to do and why. Yes, AI security is a technical issue, but it is also a governance issue.”

The white papers highlight several critical priorities for enterprises:

  • Embedding AI security into Governance, Risk, and Compliance (GRC) frameworks
  • Moving beyond model-centric controls to system-level runtime guardrails
  • Managing delegated authority in agentic AI systems
  • Combining detection with verification where certainty is required
  • Establishing measurable, independent validation practices

Together, the papers provide a practical roadmap for organizations to safely transition from AI experimentation to accountable, production-grade deployment.

Both white papers are available for download at nsslabs.com.

Cybersecurity Testing Pioneer Bob Walder Joins NSS Labs as Senior Analyst

Austin, Texas — March 17, 2026 — NSS Labs today announced that cybersecurity testing pioneer Bob Walder has joined the organization as Senior Analyst, where he will focus on independent research and advising clients on the security implications of artificial intelligence technologies.

Walder is widely recognized as the founder of The NSS Group, Europe’s first independent network security testing laboratory, which he established in 1991. In 2007, Vikram Phatak acquired the assets, establishing NSS Labs, Inc. in the United States.

Following the acquisition, Walder served as Research Director (EMEA) for Security, Privacy & Risk at Gartner, advising enterprise organizations on cybersecurity strategy and risk management.

With NSS Labs, Inc. (1.0) rapidly growing, in 2011 Phatak persuaded Walder to join NSS Labs as President and Chief Technology Officer. Walder built a team of research analysts dedicated to helping enterprise organizations navigate complex cybersecurity challenges and make informed decisions about security infrastructure investments.

In his new role as Senior Analyst at NSS Labs LLC (2.0), Walder will focus on independent research and analysis exploring how artificial intelligence is transforming the cybersecurity landscape, including both the defensive capabilities AI enables and the new attack surfaces it creates.

“Artificial intelligence is rapidly reshaping the cybersecurity landscape, creating both powerful new offensive and defensive tools as well as entirely new categories of risk,” said Vikram Phatak, CEO of NSS Labs. “Bob has spent decades helping enterprises understand complex security technologies through rigorous independent analysis. His return strengthens our ability to provide trusted insight into one of the most important technology shifts facing the industry today.”

“AI is already transforming how both attackers and defenders operate,” said Bob Walder. “Organizations need clear, independent analysis to understand how these technologies affect risk, resilience, and the security of critical infrastructure. I’m excited to contribute research that helps enterprises navigate this rapidly evolving landscape.

CYBER WISEGUYS: Cybersecurity Solutions Independent Validation with Vikram Phatak, CEO @ NSS Labs

In this episode, Greg & Danny are joined by Vikram Phatak, CEO of NSS Labs. Vikram shares insights on cybersecurity testing, the challenges of third-party validation, the evolving landscape of AI security and the delicate balance between confidentiality and transparency in cybersecurity.

Key takeaways:

  1. Ask vendors about what their solutions do and do not do
  2. Prioritize real-world testing over marketing claims
  3. Invest in independent validation and testing

InsideCybersecurity: Cyber Assessment Firm Identifies Evasion Vulnerabilities in Enterprise Firewall Products

A nonprofit cyber assessment firm found vulnerabilities in the ability of widely used enterprise firewall products to block transport and network-layer evasions commonly deployed by cyber attackers, in a report examining the effectiveness of security offerings.

“Enterprise Firewalls are constantly evolving to combat new attacker techniques and tools but sometimes that evolution takes a wrong turn. A vendor can have a near-perfect detection engine but if attackers can bypass that engine it gives them a clear path through your defenses,” CyberRatings.org CEO Vikram Phatak sad in a Nov. 5 release.

CyberRatings is a nonprofit organization conducting independent testing of cybersecurity products through its testing partner firm, NSS Labs.

CyberRatings evaluated the “security effectiveness” of seven firewall products in 55 performance tests using 3,326 exploits, 11,311 malware samples, 5,752 evasion techniques in 53 evasion categories and 6,481 false-positive samples,” according to the report.

Read the full article here.

SDxCentral: Palo Alto Networks and Fortinet Given All Clear After Firewall Hiccups

Palo Alto Networks and Fortinet have received a clean bill of health for their firewall protections, while the jury is still out on current Cisco defenses.

CyberRatings.org recommended both Palo Alto and Fortinet after new tests confirmed they had patched evasions previously discovered by the security testing firm.

In tests carried out at the start of the month by CyberRatings’ testing partner NSS Labs, researchers found they were able to bypass protection using Layer 4 TCP evasions in both Palo Alto’s PAN-OS (version 11.2.8-c537) and Fortinet’s IPS (v7.01154), as well as evading Layer 3 IP in the Palo Alto operating system.

Both firms reacted quickly, with Palo Alto developing an updated PAN-OS firmware package (PAN-OS 11.2.10-c37) and Fortinet deploying an updated IPS package (v7.01165 (33.00064) to fix the vulnerabilities.

Read the full article here.