TL;DR: Anthropic's Frontier Red Team used Claude Opus 4.6 to scan over 6,000 Firefox C++ source files in January 2026, discovering 22 security vulnerabilities — 14 rated high severity — in just two weeks. The total cost was approximately $4,000 in API credits, a fraction of what traditional penetration testing would require. Mozilla validated the findings and shipped patches to hundreds of millions of users in Firefox 148. This is AI security research operating at production scale, and it is only the beginning.
In March 2026, Anthropic published the results of a two-week security sprint that should make every CISO reconsider their vulnerability research budget. Claude Opus 4.6 — Anthropic's most capable model — scanned nearly 6,000 C++ files in the Firefox codebase, filed 112 unique bug reports, and surfaced 22 confirmed security vulnerabilities. The price tag: $4,000 in API credits. The implications for enterprise security are enormous.
What you will learn
- The Anthropic-Mozilla partnership and what triggered it
- How Claude Opus 4.6 analyzed Firefox at scale
- Breaking down the 22 vulnerabilities
- The $4,000 cost model vs. traditional pentesting
- The two working proof-of-concept exploits
- CVE-2026-2796: the JavaScript engine critical
- Mozilla's response and Firefox 148 patches
- What this means for enterprise security teams
- The bigger picture: AI as defender and attacker
- TL;DR
The Anthropic-Mozilla partnership
When Anthropic's Frontier Red Team reached out to Mozilla in late 2025, the pitch was straightforward: let us run Claude against Firefox's source code and see what happens. Mozilla, which maintains one of the most security-critical codebases on the internet, agreed to the collaboration. The result, published in Anthropic's official announcement and covered by TechCrunch and The Hacker News, marks one of the most significant demonstrations of AI-assisted security research to date.
The Frontier Red Team is Anthropic's internal unit dedicated to probing the safety and security boundaries of AI systems. In this case, the team flipped the model: instead of testing what Claude can do when misused, they tested what Claude can do when deployed as an active defense tool against a real-world codebase. Firefox was the ideal candidate — it is open source, written heavily in C++, ships to hundreds of millions of users, and has a mature and rigorous vulnerability disclosure process through Mozilla Bugzilla.
The collaboration was structured as a responsible disclosure exercise. Anthropic's team worked closely with Mozilla engineers throughout the two-week sprint, with Mozilla helping calibrate what findings were genuinely security-relevant versus general code quality issues. That iterative feedback loop proved essential: without Mozilla's domain expertise, the signal-to-noise ratio in Claude's output would have been far lower. This is a template other partnerships should study carefully.
What makes this partnership notable beyond the raw numbers is the fact that it happened at all. Major browser vendors have historically been cautious about external AI-driven code analysis — the risk of false positives flooding triage queues is real. Mozilla's willingness to engage seriously with Anthropic signals a broader shift in how the security community is beginning to think about AI as a legitimate research tool, not just a threat vector.
How Claude Opus 4.6 analyzed Firefox at scale
The mechanics of how Claude scanned Firefox's codebase reveal something important about where AI-assisted security research currently stands — and where it is headed. According to Anthropic's red team blog post, Claude Opus 4.6 was given access to Firefox's C++ source files and tasked with identifying potential security vulnerabilities. Over the two-week engagement, the model analyzed approximately 6,000 C++ files.
C++ is notoriously difficult to analyze for security vulnerabilities. Its manual memory management, pointer arithmetic, and complex object lifetimes create a class of bugs — use-after-free errors, buffer overflows, type confusions — that are invisible to compilers but catastrophic in production. Human security researchers spend years developing intuition for spotting these patterns. Claude, it turns out, has absorbed enough of that intuition from training to surface real bugs at scale.
Within twenty minutes of starting exploration, Claude identified its first significant finding: a use-after-free vulnerability in Firefox's JavaScript engine. Use-after-free bugs occur when a program continues to use a memory reference after that memory has been freed — attackers can exploit this to overwrite freed memory with malicious data and achieve arbitrary code execution. This early finding set the tone for what followed.
The model was not just pattern-matching against known vulnerability signatures. Claude reasoned about code paths, data flow, and the conditions under which memory management errors could occur. The 112 unique reports submitted to Mozilla's Bugzilla tracker reflect a genuine analytical process, not a fuzzing run or a static analysis scan. Of those 112 reports, 22 were confirmed as genuine security vulnerabilities — a signal quality of roughly 20%, which compares favorably to human penetration testers operating under time pressure on unfamiliar codebases.
Breaking down the 22 vulnerabilities
Of the 22 confirmed vulnerabilities, the severity breakdown tells an important story: 14 were classified as high severity, 7 as moderate, and 1 as low. That distribution — 64% high-severity findings — is striking. It suggests Claude was not just finding low-hanging fruit or theoretical edge cases. It was surfacing the kinds of bugs that real attackers look for.
High-severity vulnerabilities in a browser context typically mean bugs that could allow an attacker to achieve remote code execution, escalate privileges, or escape the browser sandbox. Use-after-free bugs, type confusions, and memory corruption issues dominate this category. The fact that Claude identified 14 such bugs in two weeks — in a codebase that Mozilla's own security team actively monitors — is a significant result.
Beyond the 22 security-sensitive findings, Axios reported that Anthropic's team discovered approximately 90 additional bugs classified as non-security code quality issues. Many of these have also been fixed in subsequent Firefox releases. This means Claude's total contribution to Firefox's codebase health during the engagement was well over 100 discrete issues — a productivity rate that no human team of equivalent cost could match.
The moderate and low-severity findings are also worth noting. These are bugs that may not be immediately exploitable but represent weaknesses that could be chained with other vulnerabilities to create more severe attacks. In sophisticated real-world exploits, attackers rarely rely on a single high-severity bug — they compose multiple smaller issues into a full exploitation chain. Claude's ability to surface the complete spectrum of vulnerabilities, not just the flashiest ones, is precisely what makes this capability strategically valuable.
The $4,000 cost model vs. traditional pentesting
Let's talk about the number that will keep security procurement teams up at night: $4,000. That is the total cost in API credits for the entire two-week engagement — 6,000 files analyzed, 112 reports filed, 22 confirmed vulnerabilities, and 2 working proof-of-concept exploits.
Traditional penetration testing at this scale does not exist at this price point. A qualified external penetration testing firm engaging with a major browser codebase for two weeks — with researchers who understand C++ memory management at a deep level — would cost somewhere between $50,000 and $200,000 depending on firm reputation, scope, and depth of engagement. Bug bounty programs for browsers like Firefox offer payouts of $3,000 to $10,000 per high-severity vulnerability. Twenty-two vulnerabilities through a conventional bug bounty program alone would represent $66,000 to $220,000 in researcher payouts, before factoring in program management overhead.
The $4,000 figure covers not just vulnerability discovery but the entire research process: iterating on prompts, refining what counts as a reportable finding, working through false positives, and attempting exploit development. This is not a cherry-picked cost estimate — it is the full API spend for an engagement that produced enterprise-grade security research output.
This cost dynamic will reshape how security budgets are allocated. Chief Information Security Officers who currently spend six figures annually on periodic penetration tests are now looking at a world where continuous AI-driven vulnerability scanning costs less than a single day of human penetration testing. The question is not whether this changes the industry — it is how fast.
The two working proof-of-concept exploits
Finding vulnerabilities is one thing. Exploiting them is another. Anthropic's team went further than vulnerability discovery: they evaluated whether Claude could develop working proof-of-concept exploits for the bugs it found. The results are nuanced and worth understanding carefully.
Out of several hundred exploitation attempts costing the full $4,000 in API credits, Claude succeeded in producing working exploits in exactly two cases. The Hacker News noted that Claude is substantially better at finding vulnerabilities than exploiting them — the finding-to-exploitation success rate is low, and the cost of moving from vulnerability to exploit is an order of magnitude higher than the cost of finding the vulnerability in the first place. This asymmetry is important.
From a defensive standpoint, this asymmetry is actually reassuring. The fact that Claude can find bugs at scale and low cost is good news for defenders who can use the same capability continuously. The fact that turning those bugs into working exploits remains expensive and unreliable means that the most dangerous part of the attack chain — exploitation — has not been fully automated. Yet.
From an offensive standpoint, even two working exploits from an automated system is significant. Prior to systems like Claude Opus 4.6, automated exploitation of novel vulnerabilities at this level of sophistication was effectively impossible. Security researchers have noted that the gap between vulnerability discovery and exploitation is narrowing, and that trend will continue as models improve.
CVE-2026-2796: the JavaScript engine critical
The most technically significant finding from the engagement is CVE-2026-2796, a JIT miscompilation vulnerability in Firefox's JavaScript WebAssembly component. NIST assigned it a critical CVSS score of 9.8 — the near-maximum rating reserved for vulnerabilities that are remotely exploitable, require no authentication, and have maximum impact on confidentiality, integrity, and availability.
Anthropic published a detailed technical breakdown at red.anthropic.com describing how Claude reverse-engineered its own exploit for this CVE. The mechanism: Claude's exploit achieves arbitrary read/write capabilities and code execution in Firefox's JavaScript engine by leveraging type confusion in the JIT compiler. In practical terms, this means an attacker who successfully exploited this bug in an unpatched browser could execute arbitrary code on the victim's machine.
The important caveat — and Anthropic has been careful to emphasize this — is that Claude's exploit only functions in a testing environment that deliberately removes several security mitigations present in production Firefox builds. Modern browsers layer multiple defenses: Address Space Layout Randomization (ASLR), sandboxing, Control Flow Integrity (CFI), and others. Claude's exploit, as written, does not chain through these defenses to achieve what researchers call a "full-chain" exploit. Full-chain exploits, which combine multiple vulnerabilities to escape the browser sandbox entirely, are what would cause serious real-world harm — and those remain beyond current AI capability.
Mozilla patched CVE-2026-2796 in Firefox 148.0 on February 24, 2026. The fix is now live for hundreds of millions of Firefox users worldwide.
Mozilla's response and Firefox 148 patches
Mozilla's handling of this engagement demonstrates how responsible disclosure between AI labs and software vendors should work. Mozilla's blog post describes the collaborative process: Anthropic's team filed reports through Bugzilla, Mozilla's engineers triaged and validated findings, and both teams iterated on what constituted a genuinely reportable security issue versus a general code quality concern.
This back-and-forth calibration was not trivial. When an AI model generates 112 reports, distinguishing the 22 genuine security vulnerabilities from the 90 code quality issues requires exactly the kind of domain expertise that Mozilla's security team possesses. The collaboration model — AI for breadth and speed, human experts for final validation and triage — is almost certainly the pattern that will dominate responsible AI security research over the next several years.
Firefox 148.0, released February 24, 2026, included patches for the majority of the vulnerabilities Anthropic's team discovered. According to SC Media's coverage, Mozilla has deployed these fixes to hundreds of millions of users — one of the largest simultaneous security patch deployments in browser history triggered by AI-assisted research.
The speed from discovery to patch is also notable. The engagement ran in January 2026, patches shipped in late February, and the results were publicly disclosed in early March. That roughly six-week timeline from AI discovery to public patch represents a fast turnaround for vulnerabilities of this severity. Traditional penetration testing engagements often have longer remediation cycles simply because the volume of findings is lower and prioritization less urgent.
What this means for enterprise security teams
For enterprise security teams, the Anthropic-Mozilla collaboration is both an opportunity and a warning. The opportunity is obvious: AI-assisted vulnerability research is now accessible at a price point that makes continuous scanning economically viable for organizations of all sizes. A $4,000 monthly budget for AI-driven code security analysis is within reach for any company with a software product — something that was not true when the equivalent capability required a team of specialist researchers.
The practical implication is that security teams should be piloting AI-assisted code analysis now, not in two years. The technology exists, the APIs are available, and the cost structure makes it impossible to justify waiting. Organizations that run AI security scans against their own codebases before their adversaries do will have a significant advantage. The asymmetry cuts both ways: the same Claude capability that Anthropic used to find Firefox vulnerabilities can be pointed at any C++ codebase with an API key and a budget.
The warning is more subtle but equally important. If Anthropic can spend $4,000 and find 22 high-severity vulnerabilities in Firefox in two weeks, adversaries with access to similar models can run the same playbook against enterprise software. The barrier to entry for vulnerability research has dropped dramatically. Security teams that are not continuously scanning their own code are operating on borrowed time — the question is whether they find their own vulnerabilities first or discover them through an incident.
Enterprise security architecture is also going to need to evolve. Traditional penetration test cycles — quarterly or annual engagements with external firms — were designed around the economics of human-labor security research. Those economics no longer apply. AI-driven continuous scanning requires rethinking when and how vulnerability assessments happen, how findings are triaged, and how remediation is prioritized. Security operations centers that are not integrating AI-assisted analysis into their workflows will fall behind the threat landscape.
The bigger picture: AI as defender and attacker
The Firefox engagement crystallizes a tension that has been building in the security community for several years: the same AI capabilities that enable faster, cheaper, more comprehensive defense also enable faster, cheaper, more comprehensive attack. There is no version of Claude Opus 4.6 that only finds vulnerabilities for defenders.
InfoQ's analysis frames this correctly: AI is becoming a force multiplier for both sides of the security equation simultaneously. The question is not whether AI-assisted vulnerability research will be used offensively — it already is, by sophisticated nation-state and criminal actors who have access to frontier models through both legitimate and illegitimate channels. The question is whether defenders can build institutional AI security capacity fast enough to stay ahead.
Anthropic's decision to conduct this research openly and publish the results is an important signal. Transparent AI safety research, including research that demonstrates capability limits and asymmetries like the finding-to-exploitation gap, gives the security community the information it needs to calibrate defenses appropriately. Security-through-obscurity — the idea that keeping AI capability quiet keeps the world safer — does not work when the underlying models are available to anyone with an API key.
The Register's coverage noted a key distinction that should inform enterprise strategy: Claude is significantly better at finding bugs than exploiting them. This gap exists today. It will narrow over time as models improve. Enterprise security strategy should be planning for the world where that gap has closed, not the current one where it still provides a buffer.
The Mozilla engagement also suggests a new category of partnership that will become standard: AI labs and critical software vendors running joint security research with responsible disclosure protocols built in from the start. The model works — the Firefox collaboration produced real results that improved real security for real users. As more AI labs follow Anthropic's lead and more vendors become comfortable with AI-assisted analysis, this kind of partnership will become a routine part of the software security lifecycle.
TL;DR
- Anthropic's Frontier Red Team partnered with Mozilla to run Claude Opus 4.6 against Firefox's C++ codebase in January 2026
- Over two weeks, Claude analyzed approximately 6,000 C++ files and submitted 112 unique bug reports to Mozilla Bugzilla
- 22 confirmed security vulnerabilities emerged: 14 high severity, 7 moderate, 1 low — plus 90 non-security code quality bugs
- Total cost of the entire engagement: $4,000 in API credits, compared to $50,000–$200,000 for equivalent human penetration testing
- Claude produced 2 working proof-of-concept exploits out of hundreds of attempts — it is much better at finding bugs than exploiting them
- CVE-2026-2796, a critical 9.8 CVSS JIT miscompilation in Firefox's JavaScript WebAssembly component, was the most severe finding
- Mozilla shipped patches for all confirmed vulnerabilities in Firefox 148.0 on February 24, 2026, protecting hundreds of millions of users
- The finding-to-exploitation gap still exists, but it will narrow as models improve — enterprise security teams should be treating AI-assisted scanning as a continuous practice, not a periodic one
- The same capability used for defense is available for offense: organizations that do not scan their own code with AI tools are operating behind the threat curve