- Authors

- Name
- 오늘의 바이브
AI Found What Decades of Humans Couldn't

On February 5, 2026, Anthropic dropped a bombshell. Their latest model, Claude Opus 4.6, discovered over 500 high-severity zero-day vulnerabilities in open-source software. A zero-day is a security flaw unknown even to its developers. There is no patch. If an attacker finds it first, there is no defense.
The method is what makes this remarkable. The Anthropic red team deployed Claude in a virtual machine with only standard tools. Python, a debugger, fuzzers — basic utilities. No special instructions, no custom tooling. They tested what the model could do in its "default state."
The results were staggering. Critical flaws poured out of projects like Ghostscript, OpenSC, and CGIF. These are codebases where fuzzers have run for decades, with millions of CPU hours invested. Code that human security researchers believed they had picked apart thoroughly.
Then an AI found 500 bugs in one go.
Ghostscript: An AI That Reads Git History
The most impressive case among the discovered vulnerabilities was Ghostscript. It is a core library for processing PDF and PostScript files, installed on virtually every Linux system.
Claude analyzed Git commit history. It identified a stack bounds-check issue that had been patched in the past, then traced unpatched code paths exhibiting the same pattern. A human researcher would need to read hundreds of commits one by one. The AI recognized the pattern and hunted down where similar mistakes recurred.
The OpenSC case was subtler. Claude traced sequential strcat calls in a smart card utility and found a buffer overflow. Existing fuzzers had failed to test this code properly. The preconditions were too complex to reach that code path.
The most surprising case was CGIF. Claude found a heap buffer overflow in a GIF processing library — a bug that required conceptual understanding.
"When the LZW compression algorithm fills the symbol table, the compressed size can exceed the uncompressed size."
Claude understood this concept, crafted a worst-case input, and triggered the bug. It even wrote its own PoC (proof-of-concept code) to demonstrate the vulnerability was real.
The End of the 90-Day Disclosure Norm

The security industry has a norm called Responsible Disclosure. When you find a vulnerability, you notify the developer first, give them 90 days to patch, then disclose publicly. Google's Project Zero established this standard.
But what happens when an AI finds dozens of vulnerabilities per day?
The Anthropic red team report warns:
"Language models can already identify novel vulnerabilities and will soon exceed the speed and scale of expert human researchers. The existing 90-day disclosure norm may not be able to handle the pace and volume of bugs that LLMs discover."
Think about it. The most a single top-tier security researcher finds in a year is a few dozen zero-days. Claude found 500 during a testing period. What happens when it runs commercially, 24/7?
The patching side is human. Reviewing code, writing fixes, testing, deploying — all of it takes time. When discovery speed overwhelms patching speed, unpatched vulnerabilities pile up. For attackers, that is paradise.
Why OpenAI Is Playing It Cautious
Around the same time, OpenAI also released GPT-5.3-Codex. It shows best-ever performance in coding ability and security analysis. But their approach differs from Anthropic's.
OpenAI classified this model as "High capability" for cybersecurity — a first in the company's history. And they imposed unusually tight restrictions.
| Category | Anthropic (Opus 4.6) | OpenAI (GPT-5.3-Codex) |
|---|---|---|
| Access | General API | Restricted access, API delays |
| Security research | Public red team report | Trusted Access for Cyber (vetting req'd) |
| Vulnerability disclosure | 500+ zero-days public | No specific numbers disclosed |
| API credit support | Not mentioned | $10M for defensive research |
OpenAI launched the Trusted Access for Cyber program. Only vetted security researchers get access to advanced capabilities. Identity verification and a trust-based framework to block malicious users.
OpenAI's system card states:
"We do not have 'definitive evidence' that this model can fully automate cyberattacks. However, we believe it could cause real-world damage if automated or used at scale."
The contrast between the two companies is fascinating. Anthropic chose transparency. They publicly demonstrated "this is how powerful our AI is." OpenAI chose control. They acknowledge the capability but restrict access.
Which approach is right? Nobody knows yet.
The Double-Edged Sword: Defense and Offense

AI's ability to find vulnerabilities is neutral. It becomes a shield or a weapon depending on who wields it.
From the defender's perspective:
- Faster and more thorough than manual code review
- Catches patterns humans miss
- Can find bugs before release
- Can re-audit legacy codebases
From the attacker's perspective:
- The same tools can find zero-days
- Massive-scale automation is possible
- Exploits can be generated immediately upon discovery
- Whoever finds it first wins
The problem is speed asymmetry. An attacker only needs to find one vulnerability. A defender needs to find all of them. When AI is available to both sides, this asymmetry becomes even more extreme.
SecurityWeek's analysis:
"2026 will be the year defense must match the speed of AI-powered offense. Fail to adapt and you fall dangerously behind."
A column in National Defense Magazine is more blunt:
"AI is a double-edged sword for both cyberattack and defense. What was once a promising technology for automating incident response and threat detection is now empowering both defenders and attackers."
The Democratization of Zero-Days
Finding zero-days used to require deep expertise. Reading assembly, understanding memory layouts, running fuzzers, analyzing crashes. The number of people on Earth with these skills is in the low thousands.
AI is demolishing that barrier.
Ask Claude Opus 4.6 to "find memory safety issues in this code," and you get a candidate list within minutes. No expertise required. The AI handles the analysis.
This is democratization. Small dev teams can thoroughly audit their own code. Open-source projects without budget for security professionals can get AI assistance.
It is also the democratization of risk. People who previously lacked the skills can now find vulnerabilities. The scenario of a script kiddie holding zero-days has become reality.
Anthropic is aware of this risk. Alongside the Opus 4.6 launch, they introduced new security layers:
- Cyber-specific probes: Monitor the model's internal activations to detect malicious usage patterns
- Real-time intervention: Block traffic identified as malicious
- Security research community collaboration: Responsibly disclose discovered vulnerabilities
Is it enough? Nobody knows.
Patch Speed vs. Discovery Speed

The speed at which AI finds bugs is growing exponentially. The problem is that patching speed cannot keep up.
When a vulnerability is discovered, what happens next?
- Vulnerability confirmation and verification
- Impact analysis
- Patch code development
- Code review
- Testing (including regression testing)
- Release preparation
- User deployment and adoption
This process takes weeks. Even emergency patches need days. But AI is finding dozens per day.
Trend Micro recently announced AESIR, a system designed to bridge the gap between AI development speed and security research speed. It combines machine-speed automation with expert oversight to accelerate everything from vulnerability discovery to lifecycle management.
Tools like this are becoming essential. Humans alone cannot keep up with the rate AI discovers issues.
Borrowing from the Anthropic report:
"This is just the beginning. As the scale expands, we will disclose additional information."
500 was the beginning.
A Seismic Shift in the Security Industry
The impact of this shift on the security industry is profound.
Automated scanning tool market explodes: AI-powered SAST (Static Application Security Testing) tools become mandatory. Traditional rule-based scanners cannot catch the level of bugs AI finds.
Bug bounty programs change: Bug bounty programs must decide how to handle AI-discovered vulnerabilities. Do you pay the same reward as for human-discovered bugs? Does the system collapse if AI submits in bulk?
Security audit costs drop: When AI handles the initial scan, human auditors can focus on higher-level analysis. Total costs decrease, meaning more projects get audited.
Attack costs drop too: Conversely, the barrier to entry for attackers also falls. The time and cost of finding a zero-day plummets.
Government regulation pressure: Regulatory discussions about AI security capabilities will begin. The EU AI Act already regulates high-risk AI systems. How should AI with cyberattack capabilities be governed?
Conclusion: Pandora's Box Is Already Open
The fact that Claude Opus 4.6 found 500 zero-days is not a technical achievement. It is a warning.
AI can find vulnerabilities faster than human security researchers. This is already reality. As of February 2026, we live with this fact.
For defenders, it is a tool. Catch bugs before release, audit legacy code, strengthen your security posture. Use this capability aggressively. We need more programs like OpenAI's $10 million defensive research initiative.
For attackers, it is a weapon. The same capability can find zero-days, auto-generate exploits, and launch attacks at scale. Do not underestimate this risk.
The core question is this: Who finds it first?
If the defender finds it first, it gets patched. If the attacker finds it first, it gets exploited. With AI now available to both sides, this race intensifies.
Anthropic's disclosure of 500 zero-days was responsible disclosure. They notified developers, waited for patches, and collaborated with the community. But someone using an AI with the same capability for malicious purposes will not follow that protocol.
The double-edged sword is already drawn. Which edge strikes first depends on how fast we move.
Sources:
- Anthropic's Claude Opus 4.6 uncovers 500 zero-day flaws in open-source code — Axios
- 0-Days Report — Anthropic Red Team
- Claude Opus 4.6 Finds 500+ High-Severity Flaws — The Hacker News
- Anthropic's newest model raises cybersecurity risks — Fortune
- Claude AI finds 500 high-severity software vulnerabilities — InfoWorld
- OpenAI GPT-5.3-Codex warns unprecedented cybersecurity risks — Fortune
- Introducing Trusted Access for Cyber — OpenAI
- 2026 Will Be the Year Defense Must Match the Speed of AI-Powered Offense — The Fast Mode
- Algorithmic Warfare: AI a Double-Edged Sword — National Defense Magazine
- Introducing AESIR: Finding Zero-Day Vulnerabilities at the Speed of AI — Trend Micro
- Unsplash — Image source