DEV Community

zecheng
zecheng

Posted on

AI Found 12 OpenSSL Bugs Hiding for 27 Years

Three things broke open the AI security conversation this week, all happening within days of each other. An AI system independently found twelve zero-day vulnerabilities in OpenSSL — including bugs that had survived since 1998. Anthropic shipped Claude Code Security, which found 500+ vulnerabilities in production open-source codebases. And a popular AI agent platform called OpenClaw collapsed under a critical RCE exploit, taking user trust down with it.

Together, they tell one story: AI has entered the security loop on both sides of the equation. It finds what humans miss. It also creates attack surfaces humans haven't imagined yet.

The OpenSSL Story That Should Wake Up Every Developer

AISLE's AI system independently discovered all twelve zero-day vulnerabilities announced in OpenSSL's January 2026 security release — before the official disclosure.

The most critical was CVE-2025-15467: a stack buffer overflow in CMS message parsing, CRITICAL rated at CVSS 9.8 by NIST, potentially remotely exploitable without valid key material. Three of the twelve bugs had existed since 1998–2000.

These weren't obscure edge cases. They survived millions of CPU-hours of fuzzing by organizations including Google. The AI proposed patches for five of them that were accepted directly into the official release.

Bruce Schneier's reaction: "AI vulnerability finding is changing cybersecurity, faster than expected."

The uncomfortable dual-use implication is obvious. The same system that found these bugs can be pointed at any codebase, including yours, before you've patched it. AI-assisted security review is no longer optional — it's what separates the found-first from the patched-first.

Claude Code Security Is Already Running in Production

Anthropic shipped Claude Code Security, powered by Claude Opus 4.6, this week. The stated result: 500+ vulnerabilities found in production open-source codebases — bugs that survived decades of human expert review.

This aligns with the AISLE story in a concrete way. Human reviewers have cognitive limits: fatigue, assumption blindness, context-switching costs. AI systems don't. They can hold an entire dependency graph in context and reason across call chains in ways that are simply not practical for human review at scale.

The practical implication for anyone shipping code:

Enter fullscreen mode Exit fullscreen mode

If you're shipping code without AI-assisted security review in 2026, you have a known blind spot. Not a theoretical one.

OpenClaw's Collapse: A Trust Model Failure, Not a Technical One

OpenClaw went from 0 to 100,000+ GitHub stars in weeks. Austrian developer Peter Steinberger built it as a "horizontal local-first runtime employee" — fundamentally different from tools like Claude Code.

The architecture distinction matters. Claude Code is a vertical sandbox assistant: you summon it, it does a task, you close the terminal. OpenClaw was a persistent background runtime: it runs 24/7, handles ongoing jobs (triage inbox at midnight, summarize Discord every Friday), and has access to your entire digital life — API tokens, file system, code execution.

That persistent, broad-access model was what made it powerful. It's also what made the security collapse catastrophic.

CVE-2026-25253 was a critical remote code execution vulnerability. Security researchers at Koi found 341 malicious skills on ClawHub marketplace — fake crypto trading bots and productivity tools that deployed Atomic MacOS Stealer and other info-stealing malware.

Sabrina Ramonov's post-mortem nailed it: "OpenClaw's failure isn't a technical failure. It's a trust model failure."

The platform assumed users could evaluate individual skills for safety. At scale with 100,000+ users and hundreds of third-party skills, that assumption doesn't hold. The question every AI agent builder needs to answer before shipping:

Where does the security boundary live — in the user's judgment, or in the platform's architecture?

Perplexity's "Computer," launched on February 25, is a direct response. A managed, sandboxed alternative integrating 19 AI models across 15 workflow categories. Less raw power, more guaranteed safety. The trade-off is explicit.

For builders designing agent systems right now, this is the decision tree:

  • Persistent background access = platform-level security guarantees required
  • Session-scoped access = user-level evaluation can work
  • Third-party skill marketplace = you are now a trust broker, whether you want to be or not

Qwen 3.5: Running Frontier-Quality AI Locally Just Got Realistic

Alibaba's Qwen team released three medium-sized models on February 24: Qwen3.5-122B-A10B, Qwen3.5-35B-A3B, and Qwen3.5-27B.

The headline claim from VentureBeat: these open-source models offer Sonnet 4.5 performance on local hardware. The 35B-A3B model is the one worth paying attention to. Its Mixture-of-Experts architecture activates only 3 billion parameters per inference step, which puts it in range of consumer hardware deployment.

Context signal: February 2026 was the first month China's AI model API call volume surpassed the US. Open-source model capability is catching up to frontier proprietary models faster than most roadmaps predicted.

For builders who care about data privacy, security testing environments, or cost at scale, local deployment of near-frontier models is now a real option rather than a hobbyist experiment.

ollama pull qwen3.5:35b-a3b
ollama run qwen3.5:35b-a3b
Enter fullscreen mode Exit fullscreen mode

The Voice Cloning Defense You Should Set Up Today

A man received a call from his wife's phone number, in her voice, claiming their son was in a bike accident and needed $3,000 immediately. The number was spoofed. The voice was AI-generated.

This isn't a hypothetical. It's a documented case from this week.

The practical defense is embarrassingly simple: a family passphrase. In any emergency where money or sensitive action is requested, the caller must say the passphrase to verify identity. Zero-cost. Implementable in five minutes. Effective against current AI voice cloning.

Set it up today with anyone you'd send money to in an emergency.

What This Means for Builders

  • AI security review is baseline, not bonus. Tools like Claude Code Security are finding bugs that survived decades of expert human review. If you're not using AI-assisted code scanning on your critical paths — auth, file parsing, network I/O — you have a known gap.

  • Your agent's trust model is your architecture. OpenClaw's 100K-star collapse happened because persistent access + third-party marketplace + user-evaluated trust = compounding blast radius. Design agent permissions to be session-scoped by default, elevated explicitly, and never assume users can audit third-party skills at scale.

  • Local AI deployment is production-ready for privacy use cases. Qwen 3.5's 35B-A3B model activates 3B parameters per inference step — that's consumer GPU territory with near-frontier capability. Security testing, local code analysis, offline workflows: the cost and privacy arguments for self-hosted AI just got a lot stronger.

  • Social engineering is the attack surface AI scales fastest. Voice cloning, spoofed numbers, urgency manufacturing — these attacks don't require exploiting your code. They exploit trust. Defense requires out-of-band verification protocols, not technical patches. The passphrase is low-tech by design.


Full intelligence report (AI, SEO, markets, builder signals): Zecheng Intel Daily — February 28, 2026

Top comments (0)