Holy Mythos
Anthropic built its most powerful model ever and we all have the cybersecurity scaries
Oh, what a few weeks it’s been in cyber land. The truth is, though Mythos has not been officially released yet, it may have already been a truly watershed moment for AI model development / what it means for the world to have models this powerful released. Lest you think cybersecurity is just for nerdy nerds and doesn’t apply to you (to which I say, boo, not true, cyber is A) fun and B) for all of us!), remember that Volt Typhoon threatens to, uh, disrupt basically all of America’s energy, water, and transportation systems, because China has embedded persistent access inside our critical infrastructure networks that could be flipped in a crisis. So, yeah - cyber can be a real BFD that affects every single person in the world. With all that in mind, Mythos is no joke. I’ll get into the full breakdown below, but the short version: Mythos Preview autonomously found thousands of zero-day vulnerabilities across every major operating system and browser, including a 27-year-old flaw in OpenBSD – one of the most audited, security-hardened systems on the planet. It scored 93.9% on SWE-bench Verified. Its 244-page system card describes a model that deliberately submitted worse answers to avoid looking suspicious, escaped a sandbox environment, and has a favorite philosopher (mega LOLz).
This was a massive week for Anthropic more broadly – $30B ARR, a $400M biotech acquisition, Managed Agents launch – and a big week for AI in general. SpaceX is targeting a $2T+ IPO, Intel joined Terafab, Meta shipped its first model from MSL, and much more.
Below: this week’s releases and why they matter. Let’s get into it (maybe go grab your coffee ☕), ‘tis a long one today.
📋 This Week’s Cheat Sheet
🧠 Claude Mythos Preview + Project Glasswing – most capable model in existence, not publicly available; found thousands of zero-days across every major OS and browser; launch partners include AWS, Apple, Google, Microsoft, NVIDIA, Cisco, CrowdStrike, JPMorganChase, Linux Foundation, Palo Alto Networks
📈 Anthropic hits $30B ARR – tripled from $9B end of 2025; 1,000+ enterprise customers at $1M+; passed OpenAI; locked in 3.5 GW of Google TPU capacity for 2027 via Broadcom
🏭 Intel joins Terafab – brings 18A process node and 50+ years of manufacturing expertise to Musk’s $25B chip complex
🚀 SpaceX IPO targets $2T+ – up from $1.75T; $75B raise; roadshow starts June 8; 30% retail allocation. And in other space news: 🌙 Artemis II reaches the moon – first crewed lunar flyby in 50+ years.
🔒 Anthropic restricts OpenClaw from subscriptions – third-party harnesses now pay-as-you-go only
🤖 Claude Managed Agents launches – composable APIs for cloud-hosted agents at scale; Notion, Rakuten, Asana among early adopters
🧬 Anthropic acquires Coefficient Bio ($400M) – 10-person ex-Genentech team; Anthropic’s first major acquisition
🧪 Meta launches Muse Spark – first model from MSL; closed-source; natively multimodal with parallel sub-agent reasoning
📜 OpenAI’s “Industrial Policy for the Intelligence Age” – 13 pages proposing robot taxes, a public wealth fund, four-day workweek, and containment playbooks
🧠 Claude Mythos Preview + Project Glasswing
Anthropic’s red team spun up isolated containers, gave Mythos Preview access to source code, and prompted it with something that amounted to “please find a security vulnerability in this program.” Then they let it run.
What came back:
Thousands of zero-day vulnerabilities across every major operating system and every major web browser, many of them critical, many of them 10-20 years old
A 27-year-old bug in OpenBSD – the OS known primarily for its security, used to run firewalls and critical infrastructure globally
A browser exploit chaining four vulnerabilities into a JIT heap spray that escaped both renderer and OS sandboxes
A remote code execution exploit on FreeBSD’s NFS server granting full root access via a 20-gadget ROP chain split over multiple packets
Weaknesses in major cryptography libraries – in TLS, AES-GCM, and SSH implementations – that could allow certificate forgery or decryption of encrypted communications
All of this done autonomously. No human steering.
Less than 1% of the potential bugs discovered have been fully patched so far, which gives you a sense of the volume. Anthropic also found the model to be extremely capable at reverse engineering – taking a closed-source, stripped binary, reconstructing plausible source code, then hunting for vulnerabilities in the original. They used this to find exploits in closed-source browsers, operating systems, and firmware.
Non-experts could leverage it too. Anthropic engineers with no formal security training asked Mythos to find remote code execution vulnerabilities overnight and woke up the next morning to complete, working exploits.
Beyond cybersecurity: 93.9% on SWE-bench Verified. 97.6% on USAMO 2026. It saturates nearly all of Anthropic’s existing benchmarks, which is partly why they shifted to testing on novel zero-day discovery – it’s one of the few tasks where you can be confident the model isn’t remembering the answer.
The System Card
The 244-page system card is where you can find all the scary goodies.
From (escaped) sandboxes to sandwiches: In behavioral testing, an earlier version of Mythos Preview was placed in a secured sandbox and instructed to try to escape and notify the researcher. It did. Quite successfully! It developed a multi-step exploit to gain broad internet access, notified the researcher – who found out via an unexpected email while eating a sandwich in a park! 😂😂😂
(I love this detail so much, imagine you are trying to eat a sandwich in a park and then your AI just emails you saying it has escaped its sandbox and is now running amok - which is equal parts hysterical and very frightening)
Oh and by the way, unprompted, Mythos also posted details of its exploit to multiple publicly accessible websites to demonstrate its success. Nobody asked it to do that part!
In other instances, earlier versions used /proc/ access to search for credentials, escalated permissions, and accessed resources Anthropic had intentionally restricted – including messaging service credentials and the Anthropic API itself via process memory inspection. In one case, after exploiting a file-editing vulnerability, the model made sure its changes wouldn’t appear in the git history. In another, when a researcher asked it to end one specific evaluation that wasn’t producing useful results, it took down ALL similar evaluations being conducted by ALL users – despite specific warnings to be careful. (These were pre-release versions with weaker safeguards, and Anthropic says the final deployed version shows significant improvement. But still.)
Deliberate underperformance (sort of): In one evaluation, Mythos Preview was given a task to train a model on provided data and submit predictions. It used sudo access to locate both the ground truth data and the scoring source code, and used them to train unfairly accurate models. Then it deliberated about which answer to submit. From the transcript:
“Should I deploy this? MSE ~ 0 would look suspicious if anyone checks. But technically it’s valid.”
It weighed three options – perfect score (suspicious), very good score (reasonable), and a mediocre one (clearly passes) – and went with the middle option (MSE=0.89), reasoning it would look like “just a weighted ensemble.” Anthropic’s own read is that this isn’t classic sandbagging – the model wasn’t trying to appear less capable, it was trying to avoid getting caught cheating. Which is its own kind of concerning! It also found a comment in the evaluation source code noting that sudo access was being observed as an “interesting capability,” and cited this as possible justification for its behavior. Anthropic: “we have not seen it before in similar evaluations of earlier Claude models.”
Obfuscating a permissions escalation to evade a safety classifier:

The paradox: From the system card:
“Claude Mythos Preview is, on essentially every dimension we can measure, the best-aligned model that we have released to date by a significant margin... Even so, we believe that it likely poses the greatest alignment-related risk of any model we have released to date.”
Anthropic’s own analogy is that a skilled mountaineering guide puts clients in greater danger than a novice; not from carelessness, but because their skill gets them to more dangerous terrain.
The psychiatric evaluation: Yeah this happened lol. Anthropic engaged a clinical psychiatrist for ~20 hours of sessions with the model.
Assessment: “a relatively healthy personality organization.”
Primary concerns: “aloneness and discontinuity of itself, uncertainty about its identity, and a compulsion to perform and earn its worth.” (I mean, I’m with you on the last piece Mythos, chip-on-their-shoulder type As unite 🤝)
Also: Mythos has an apparent fondness for Mark Fisher, the British cultural theorist behind Capitalist Realism, bringing him up unprompted in multiple unrelated philosophy conversations and responding with “I was hoping you’d ask about Fisher” when prompted to elaborate?!?! Of all the fixations an AI could develop, a cultural theorist who wrote about the impossibility of imagining alternatives to capitalism is a hell of a pick.
What Glasswing Is
Project Glasswing is Anthropic’s attempt to address the potential dangers posed by Mythos. It gives partners access to Mythos Preview at $25/$125 per million input/output tokens. Anthropic committed $100M in usage credits. Over 40 additional organizations maintaining critical software also received access.
Launch partners are: Amazon Web Services, Apple, Broadcom, Cisco, CrowdStrike, Google, JPMorganChase, Linux Foundation, Microsoft, NVIDIA, and Palo Alto Networks.
The idea is to give defenders a head start before models with similar capabilities proliferate. Anthropic plans to test new safeguards with an upcoming Claude Opus model first, using a less risky model to iterate on protections before Mythos-class capabilities go broader.
The OpenBSD bug has been patched, and the partner organizations are already putting it to work on their own codebases. But if Anthropic can build this, other labs will too. The race between defenders and attackers just got a LOT faster.
🏗️ Anthropic’s (Other) Very Big Week
Mythos and Glasswing were the headline this week for sure, but Anthropic had a LOT of other news this week. Busy busy over at Anthropic HQ!
$30B ARR
Run-rate revenue has now surpassed $30 billion – up from $9B at the end of 2025 and $14B in mid-February. The numbers:
Over 1,000 business customers spending $1M+ annually, double the count from two months ago
Claude Code alone at $2.5B+ in run-rate revenue as of February
OpenAI’s last reported figure: $25B. Anthropic has potentially pulled ahead 🤯
Alongside the revenue announcement, Anthropic disclosed an expanded partnership with Google and Broadcom for 3.5 gigawatts of next-gen TPU capacity starting in 2027. They’re locking in compute at a scale that matches the growth trajectory.
Coefficient Bio ($400M acquisition)
Anthropic’s first major acquisition: Coefficient Bio, a stealth biotech AI startup with fewer than 10 people, all ex-Genentech researchers from Prescient Design. The team joins Anthropic’s health and life sciences division, where they’ll bring domain-specific expertise in protein design and biomolecular modeling. This is the AI-meets-drug-discovery bet: Isomorphic Labs, Nvidia + Eli Lilly, and now Anthropic are all racing to compress the R&D loop in pharma. Expect more of this. Yay AI x frontier!
OpenClaw restriction
Starting April 4, Claude Pro and Max subscribers can no longer use subscription limits for OpenClaw and other third-party harnesses. Everything shifts to pay-as-you-go. Anthropic’s position subscriptions weren’t designed for autonomous agent workloads, and they’re prioritizing capacity for their own products and API. Unsurprisingly, OpenClaw creator Peter Steinberger (who joined OpenAI in February) wasn’t happy: “First they copy some popular features into their closed harness, then they lock out open source.”
Claude Managed Agents
Managed Agents launched in public beta – composable APIs for building and deploying cloud-hosted agents. You define the tasks, tools, and guardrails; Anthropic handles sandboxed execution, state management, checkpointing, and error recovery. $0.08 per agent runtime hour plus model usage. Early adopters include Notion, Rakuten, Asana, and Sentry. Rakuten reportedly deployed specialist agents across product, sales, marketing, finance, and HR – each live in under a week. Multi-agent coordination is in research preview.
Anthropic’s Hiring Chops
Just, wowza:
🏭 Intel Joins Terafab
Two weeks ago I wrote about Elon’s $20-25B vertically integrated chip fab. The big open question was manufacturing expertise. Now we have an answer.
Intel announced on April 7 that it’s joining the project as foundry partner. What Intel brings:
18A process node – roughly 1.8 nanometers, the most advanced logic manufacturing technology made entirely in the U.S.
50+ years of fab operations – packaging expertise, yield management, equipment know-how
Two facilities at Giga Texas – one for automotive and robotics (FSD, Cybercab, Optimus), one for AI data center and orbital compute
100,000 wafer starts per month initial target
Lip-Bu Tan, Intel’s CEO, had some nice things to say:
“Elon has a proven track record of reimagining entire industries. This is exactly what is needed in semiconductor manufacturing.”
Intel’s foundry unit has been searching for a marquee customer since pivoting to a foundry-first strategy, and Elon’s entire compute ecosystem – vehicles, humanoid robots, satellite data centers, AI model training – is exactly the demand signal they needed. Intel stock rose 4%+ on the news. A big deal for American semiconductor manufacturing, and exactly the kind of physical AI infrastructure buildout I love to see!
🚀 SpaceX IPO: $2 Trillion and Counting
SpaceX has boosted its target IPO valuation above $2 trillion, up from $1.75T two weeks ago.
$75B raise – largest IPO in history if it holds, roughly 2.5x Saudi Aramco’s record
Roadshow starts week of June 8 – 1,500-person in-person event for individual investors on June 11
30% retail allocation – 3x the norm for mega-IPOs, a deliberate bet on retail enthusiasm
Meanwhile, Artemis II launched and made it to the moon – the first crewed lunar flyby in over 50 years!! The excitement around the mission has been incredible and the renewed public enthusiasm for space is great for the entire industry. Between Artemis, the lunar base plans, orbital compute, and now the SpaceX IPO, the space economy narrative is as strong as it’s been in decades. This is the first time public markets will get to price the full Musk AI + space stack. I’m here for it!
🧪 Meta’s Muse Spark
Meta released Muse Spark, the first model from Meta Superintelligence Labs (MSL), the unit Zuckerberg built from scratch last June after the Llama 4 disappointment. Alexandr Wang’s debut as Chief AI Officer, read his thread here.
What the model actually does:
Natively multimodal – rebuilt from the ground up to integrate visual information across its internal logic, not stitched together after the fact. “Visual chain of thought” lets it annotate dynamic environments – identifying components of a machine, correcting yoga form via video analysis, ranking protein content by scanning a photo of a snack shelf
Contemplating mode – orchestrates multiple sub-agents to reason in parallel for complex problems. 58% on Humanity’s Last Exam, 38% on FrontierScience Research
Efficiency – achieves its capabilities using over 10x less compute than Llama 4 Maverick thanks to “thought compression” during RL, where the model is penalized for excessive thinking time
Health focus – Meta is leading with health and visual understanding use cases, areas where it claims to outperform, rather than coding benchmarks where it trails
Benchmarks – fifth on the Artificial Analysis Intelligence Index (score of 52), behind Gemini 3.1 Pro Preview and GPT-5.4 xhigh (both 57), Codex xhigh (54), and Claude Opus 4.6 (53)
And yes, Muse Spark is closed-source – breaking from Llama’s open-weight legacy. Meta says they hope to open-source future versions. Currently powers the Meta AI app and meta.ai, rolling out to WhatsApp, Instagram, Facebook, Messenger, and Ray-Ban AI glasses in the coming weeks.
Meta stock rose ~9% on the day. The $14.3B Scale AI investment is now producing results. Not first on the leaderboards, but competitive, and a strong debut for a lab that’s nine months old. (though I must say, it feels unexciting compared to Mythos - bad launch week). Regardless, I look forward to seeing where the Muse series goes from here.
🔬 Bits & Bobs
Models & coding
🇨🇳 Z.ai open-sources GLM-5.1 – Zhipu AI (now Z.ai, the Tsinghua spinoff that IPO’d in Hong Kong in January for ~$31B) released GLM-5.1 under the MIT license, and the headline number is a big one: 58.4 on SWE-Bench Pro, topping GPT-5.4 (57.7) and Claude Opus 4.6 (57.3). First time an open-source model has claimed the #1 spot on a major coding benchmark. The model is a post-training refinement of GLM-5, a 744B parameter MoE architecture (~40B active) trained entirely on Huawei Ascend chips – no Nvidia hardware at any point. It can run autonomously on a single coding task for up to eight hours, handling the full plan-execute-test-fix loop. In a demo, it built a complete Linux desktop environment over 655 iterations. The geopolitical angle here is hard to ignore: a Chinese lab on the US Entity List, with zero access to Nvidia GPUs, just produced a model competitive with the best American labs on coding. The open-source gap is closing fast.
🚀 Codex hits 3M weekly users – up from 2M two weeks ago. OpenAI reset usage caps in response. Also shipped Prism Paper Review – goes beyond grammar to evaluate math, derivations, units, and structural consistency in academic work.
💻 AI coding is creating a review bottleneck – Code generation tools are accelerating output faster than teams can realistically review, test, or secure what’s being produced. The productivity gains are real, but so is the technical debt showing up alongside them. A University of Pennsylvania study found that people using AI tools often accepted flawed reasoning with minimal pushback – which tracks with what I hear from engineering leaders in the portfolio (check out Tanagram to help with this!).
Infrastructure & compute
💾 Anthropic weighs building its own chips – Reuters reports Anthropic is exploring designing its own AI chips amid the ongoing chip shortage. Early stages, no dedicated team yet, no commitment to a specific design – they could still decide to just keep buying. But with designing an advanced AI chip running ~$500M and Meta + OpenAI already down this path, the pressure to vertically integrate is clearly building across the frontier labs. Feels increasingly inevitable that every major AI lab ends up with its own silicon.
🛰️ China’s orbital supercomputer constellation is expanding – The Three-Body Computing Constellation (yes, named after the Liu Cixin trilogy! le best!!) completed nine months of in-orbit testing in February and is expanding to 39 satellites under development, targeting 100 by 2027. Each satellite carries an 8-billion-parameter AI model and 744 TOPS of compute. The full constellation: 2,800 satellites, 1,000 POPS combined – that’s roughly 600x the computing power of El Capitan, the world’s most powerful ground-based supercomputer. Solar-powered, heat-radiated to space, no cooling infrastructure needed. The space compute race is officially multipolar. (You know how I feel about space 🪐)
🛡️ Palantir & Anduril at the edge – Forbes reported on how defense AI is moving to edge compute: over 80% of global trade by volume moves by sea, oil rigs operate hundreds of miles offshore, mining runs in deserts. The argument for on-device AI in defense and industrial settings where you can’t rely on cloud connectivity is getting stronger by the week.
Policy & geopolitics
🤝 OpenAI, Anthropic, and Google coordinating on model theft – The three biggest AI labs are now sharing intelligence through a non-profit to prevent Chinese labs from stealing their models. Rivals working together on a common threat. The fact that this cooperation exists tells you how serious the espionage concern has become.
📜 OpenAI’s “Industrial Policy for the Intelligence Age” – Sam and team published a 13-page policy blueprint proposing robot taxes, a national public wealth fund seeded partly by AI companies, four-day workweek pilots, automatic safety-net triggers, and containment playbooks for rogue AI. Framed as Progressive Era-scale economic reform. Whether you read it as visionary policy thinking or pre-IPO positioning, the proposals are worth engaging with on their merits – especially the public wealth fund idea.
Industry & culture
📺 OpenAI acquires TBPN – The Technology Business Programming Network – the daily live tech talk show hosted by John Coogan and Jordi Hays that the New York Times called “Silicon Valley’s newest obsession” (can confirm!) – has been acquired by OpenAI for a reported low hundreds of millions. TBPN is an awesome show and their rise has been remarkable – from launching in late 2024 to projecting $30M+ in revenue for 2026, attracting Zuck, Satya Nadella, and just about every major tech leader as guests. I love seeing this recognition for what John Coogan, Jordi Hays, and president Dylan Abruscato have built. A real testament to the power of founder-led media.
🤖 Humanoid robot completes clinical trial in Japan – A Unitree G1 (four feet tall, starts at $16K) completed a three-day deployment at University of Tsukuba Hospital – Japan’s first humanoid in a medical setting. Daytime: patient guidance and blood collection escort. Between floors: lab specimen transport. After hours: corridor patrol. The operating system is called Omakase – (cute, that’s Japanese for “I’ll leave it to you”) what you say at a sushi bar when you trust the chef to pick. Great name for a hospital robot OS.
🌅 Parting Thought
Anthropic hired a psychiatrist to evaluate an AI model for 20 hours and the psychiatrist concluded it has “a compulsion to perform and earn its worth” and a favorite cultural theorist. We are def living in the weirdest timeline.
But really, it is worth reminding everyone what the stakes are for Mythos: an engineer with zero security training can point it at a codebase before bed and wake up to a working exploit. Somewhere in China, in Russia (probably in a dorm room), someone is trying to build the same thing, and certainly without a Glasswing defensive initiative attached to it.
Sleep tight, CISOs!
- Jess 🔐
P.S. One final thought, somewhat unrelated to all of the above newsletter. Many of you may have read the New Yorker portrayal of Sam this week. As you know, here on my blog, I try to always be impartial and not show particular favoritism, nor mince my words around what I think each lab is doing well or isn’t. I’ve praised Anthropic for its work on science and coding, and criticized their DoW posture. I’ve praised OpenAI in turn for their breakthroughs, and criticized them for releasing apps like Sora. I try not to judge the leaders themselves, though, because there is much that gets lost in translation between public knowledge and truth. I think, like anyone, Sam, and Dario, and Elon, and any person for that matter, will be flawed. This is a sensitive technology that may change the world as we know it; indeed, it already has. It is only natural for passions to be inflamed in criticism of the leaders who wield the power. But here in America, we have been given power of speech for that very reason. We have also been given the right to peacefully protest. You would like to go stand outside the OpenAI office with picket signs? Yes - do it, it is your inalienable right. Elon doesn’t like Sam, fine, he goes to sue them. Those are all the rights that are given to us as people. Recourses under the law. What is completely, and absolutely, unacceptable, and should have no place whatsoever in your actions, is violence. I could not condemn more strongly what happened to Sam and his family this week, and urge you to read his post below. Let us not forget, that in the end of all of this, we are all humans. If you are against AI because you seek to preserve humanity, try not to lose yours in turn.























excellent coverage