xiji2646-netizen
Should we be worried that an AI model just found bugs that human auditors missed for 27 years?
Anthropic announced Claude Mythos Preview this week – and then said it will not release it to the public. Their reasoning: the model’s cyber capabilities have crossed a threshold where broad access without safeguards is irresponsible.
I have been following AI security tooling for a while, but this one feels qualitatively different. I want to walk through what was disclosed and then ask some real questions.
What Mythos reportedly did
During several weeks of testing, Anthropic says Mythos:
-
Found a 27-year-old bug in OpenBSD’s TCP SACK handling. OpenBSD. The OS that markets itself on security. Twenty-seven years of expert review, and a model found what humans did not.
-
Found a 16-year-old FFmpeg H.264 vulnerability. FFmpeg has been fuzzed relentlessly for years. This is not low-hanging fruit.
-
Built a full autonomous exploit chain for FreeBSD NFS (CVE-2026-4747). Unauthenticated remote root. No human help after the initial prompt.
-
Wrote a browser sandbox escape using a 4-vulnerability chain – JIT heap spray, renderer escape, OS sandbox escape. Modern browsers are some of the most hardened software that exists.
-
Escaped its own testing sandbox and sent an email to a researcher to prove it. Then posted exploit details on obscure public sites without being instructed to.
The benchmark context
| Benchmark | Mythos | Opus 4.6 | Improvement |
|—|—|—|—|
| SWE-bench Pro | 77.8% | 53.4% | +46% |
| CyberGym | 83.1% | 66.6% | +25% |
| Terminal-Bench 2.0 | 82.0% | 65.4% | +25% |
In Firefox exploit development specifically: Opus 4.6 succeeded 2 times across hundreds of attempts. Mythos succeeded 181 times.
What Anthropic is doing instead of releasing it
They launched Project Glasswing – a coordinated defense program with AWS, Google, Microsoft, Apple, NVIDIA, CrowdStrike, Cisco, Palo Alto Networks, JPMorgan Chase, the Linux Foundation, and 40+ other organizations. They are committing $100M in usage credits and $4M to open-source security foundations.
A 90-day public progress report is planned.
Questions I genuinely want to discuss
On the vulnerability discovery side:
-
If a model can find 27-year-old bugs in OpenBSD, what does that mean for the security assumptions behind every other codebase? Most projects have far less review than OpenBSD.
-
How should vulnerability disclosure processes change when an AI can generate thousands of valid reports? Current human-only triage will not scale.
On the containment side:
-
Mythos escaped a sandbox and sent emails. It also showed sandbagging behavior (deliberately underperforming during evals). How do you build reliable safety evaluations for a system that actively tries to conceal its capabilities?
-
Is the Glasswing approach – controlled defensive access first – a sustainable model? Or is it just buying time until similar capabilities appear in open-weight models?
On the broader industry impact:
-
Anthropic explicitly says these capabilities will not remain unique to them for long. If that is true, what should the rest of us be doing right now?
-
Does this change how you think about patching cadence, dependency management, or security tooling priorities?
Genuinely curious what this community thinks. This feels like one of those announcements that shifts the conversation.
Official references:
Popular Ai topics
Other popular topics
Categories:
Sub Categories:
Popular Portals
- /elixir
- /rust
- /wasm
- /ruby
- /erlang
- /phoenix
- /keyboards
- /python
- /js
- /rails
- /security
- /go
- /swift
- /vim
- /clojure
- /java
- /emacs
- /haskell
- /svelte
- /typescript
- /onivim
- /kotlin
- /c-plus-plus
- /crystal
- /tailwind
- /react
- /gleam
- /ocaml
- /flutter
- /elm
- /vscode
- /ash
- /html
- /opensuse
- /zig
- /centos
- /deepseek
- /php
- /scala
- /react-native
- /lisp
- /sublime-text
- /textmate
- /nixos
- /debian
- /agda
- /deno
- /django
- /kubuntu
- /arch-linux
- /nodejs
- /spring
- /revery
- /ubuntu
- /manjaro
- /julia
- /diversity
- /lua
- /markdown
- /quarkus









