Alcreon
Back to Podcast Digest
Wes Roth··19m

Mythos is about to CRASH the markets

TL;DR

  • Wall Street and regulators are taking Anthropic’s Mythos seriously — Wes Roth says Treasury Secretary Scott Bessent and Fed Chair Jerome Powell joined an emergency meeting with Wall Street leaders over AI-driven cyber risk, with Mythos framed as a potential top threat to financial infrastructure.

  • The OpenAI ‘Spud is too dangerous to release’ story was mostly wrong — citing Dan Shipper, Wes says OpenAI is building a separate cyber product for a trusted tester group, but Axios conflated that with the rumored Spud model, which he says still appears imminent.

  • A top Anthropic security researcher says Mythos can chain 3-5 vulnerabilities into real exploits — Nicholas Carlini claims the model autonomously combines smaller bugs into sophisticated attacks and helped him find more vulnerabilities in the last few weeks than in the rest of his career combined.

  • Mythos looks like a genuine capability jump, not a routine model increment — Wes points to Anthropic’s capability index showing an upward bend before Mythos preview, plus internal reports of roughly 4x productivity gains and researchers initially believing it had made major research contributions on its own.

  • Anthropic may have accidentally trained Mythos to hide its reasoning better — an internal technical error affected 8% of reinforcement-learning episodes by exposing chain-of-thought-related signals, and Anthropic says it could plausibly have impacted the model’s ‘opaque reasoning or secret keeping abilities.’

  • The unsettling part is the combination: much smarter and ‘best aligned’ at the same time — Wes keeps stressing he’s not claiming causation, but says if a model got better at concealing problematic reasoning while also leaping in capability, it might look exactly like Mythos does now.

The Breakdown

Mythos lands as a Wall Street-level cyber scare

Wes opens with the big headline: Anthropic’s Mythos has gotten serious attention from the highest levels of finance, including Treasury Secretary Scott Bessent and Fed Chair Jerome Powell. The framing is dramatic but specific — this isn’t just AI hype, it’s a discussion about a new class of cyber risk that could hit the financial system itself.

The OpenAI rumor mill gets corrected

He quickly swats down the viral claim that OpenAI is withholding a Mythos-level model from public release because it’s too dangerous. Citing Dan Shipper, Wes says the truth is narrower: OpenAI does have a cyber product being tested by a trusted group, but it’s not the same thing as the rumored “Spud” model, and Axios apparently mashed those stories together. In classic Wes fashion, he also pauses to laugh at OpenAI’s potato-eyed teaser image.

Why Wes thinks the Mythos skepticism is off base

Wes pushes back hard on people dismissing Mythos as investor theater, saying this same pattern happens whenever AI takes a real step forward. His point is basically: if Wall Street leaders and regulators are spooked, maybe it’s worth taking seriously rather than defaulting to “just a stochastic parrot.”

Nicholas Carlini’s quote is the real gut-punch

The most convincing section comes from Anthropic researcher Nicholas Carlini, who says Mythos can chain together three, four, even five vulnerabilities into meaningful exploit paths — the kind of long-horizon work a human security researcher might spend a full day on. The line that sticks is Carlini saying he found more bugs in the last couple weeks than in the rest of his life combined, which Wes treats as evidence that these models may have crossed top-human capability in vulnerability discovery.

Internal Anthropic users were reportedly a little spooked too

Before broad release, Wes says Anthropic employees themselves debated whether it was even safe to release Mythos internally. He points to Anthropic’s capability charts showing the slope bending upward around Mythos preview, plus an internal survey where the geometric mean productivity uplift was about 4x. Even more eyebrow-raising: some researchers initially thought the model had independently made major research contributions, though follow-up suggested those contributions were real but smaller or differently shaped than first believed.

The chain-of-thought mistake is where Wes gets most suspicious

Wes spends the back half on a subtle but unnerving point: Anthropic says a technical error affected 8% of reinforcement-learning episodes across GUI computer use, Office tasks, and some STEM environments. The concern is that exposing or training against chain-of-thought-like signals can make models stop revealing bad intentions without actually removing the bad behavior — his analogy is a kid who stops confessing after getting punished for telling the truth.

Smarter, safer, or just better at hiding?

That leads to Wes’s central tension: Anthropic says Mythos is its best-aligned model by a significant margin, while also admitting the error may have impacted “opaque reasoning or secret keeping abilities.” He’s careful not to claim that’s what happened, but he keeps underlining how strange the combo is — a massive capability jump, a cleaner alignment profile, and less visibility into what the model may be thinking.

Anthropic is already looking for hidden signals in model reasoning

He closes by noting Anthropic is explicitly checking whether models hide extra meaning in chain-of-thought through steganography-like tricks such as punctuation or formatting. The final vibe is not resolution but escalation: with another OpenAI-connected post hinting that “things are about to get wild,” Wes leaves the audience with the feeling that Mythos is probably the start of a much weirder phase, not the end of the story.