AI Research

Anthropic's Jack Clark Puts 60% Odds on Self-Training AI by 2028

The Anthropic co-founder calls it a reluctant view: a 60% chance frontier AI trains its own successor by end of 2028.

Oliver Senti
Oliver SentiSenior AI Editor
May 5, 20263 min read
Share:
Two server racks reflecting infinitely down a darkened data center corridor, suggesting recursion

Anthropic co-founder Jack Clark posted an essay Monday putting roughly 60% odds on automated AI R&D by the end of 2028, the threshold he defines as a frontier model autonomously training its own successor. Push him on a 2027 probability, and the number drops to 30%.

He calls it a reluctant view. "I don't know how to wrap my head around it," Clark writes near the top of the latest Import AI issue, before spending the next several thousand words building a case that makes the conclusion harder to dismiss.

What counts as the threshold

Clark isn't predicting a frontier model will train another frontier model in 2026. Frontier training runs are expensive and human-intensive, and he says so directly. The narrower forecast is a proof-of-concept at the non-frontier stage, where one model end-to-end trains a successor within a year or two. Full frontier-scale automation comes later, by late 2028.

The framing throughout is unusually personal for a benchmarks essay. "Staring into the black hole," reads one section header. We may be about to witness a profound change in how the world works, Clark writes near the end. If a different essayist filed those lines you'd roll your eyes. The fact that the person writing them runs policy at one of the labs actually building this is what makes them harder to wave away.

The evidence stack

Most of the essay is benchmark accounting. SWE-Bench has effectively saturated: Clark cites Claude Mythos Preview hitting 93.9%, against a roughly 2% score for Claude 2 in late 2023. METR's time-horizon plot, which tracks how long humans take to do tasks AI completes with 50% reliability, has gone from 30 seconds with GPT-3.5 to roughly 12 hours with Opus 4.6 in about four years.

Then the AI-on-AI benchmarks. The CORE-Bench paper, which tests whether models can reproduce research papers from their repos, was declared solved in December 2025 with Opus 4.5 hitting 95.5%. On Anthropic's internal CPU training-speedup task, Claude Mythos Preview achieved a 52x speedup in April. A human researcher, per Anthropic, takes 4 to 8 hours to hit 4x on the same task.

The kernel-design and PostTrainBench progress is messier. Frontier models can post-train smaller models to roughly half the uplift human researchers achieve. Not parity. Not nothing.

A skeptic reads the same data

Most of these benchmarks come from labs that benefit from accelerating AI timelines, and several come from Anthropic itself. The CPU training-speedup figure is an Anthropic-reported result on an Anthropic-designed task, and the human baseline of "4 to 8 hours" lands without methodology behind it. Clark also leans on Anthropic's alignment research proof-of-concept as evidence, which it is, but Anthropic publishing about Anthropic isn't independent confirmation.

Outside the company, Redwood Research's Ryan Greenblatt recently doubled his own estimate, from 15% to 30%, on full AI R&D automation by end of 2028. His earlier skeptical analysis put the median for "powerful AI" closer to 2031. Even after the update, Greenblatt's number is half of Clark's.

Clark acknowledges the gap implicitly. He notes creativity and "heterodox insights" remain the part current systems struggle with, then points out that AlphaGo's Move 37 was a decade ago and hasn't been replaced by anything obviously more impressive since. That's a strange kind of reassurance to include in an essay arguing the opposite direction.

What's next

Clark says he'll spend most of 2026 working through the implications. The nearest concrete checkpoint is OpenAI's stated target, flagged in Altman's post, of an automated AI research intern by September 2026. Whether that intern ships, and what it can actually do if it does, will say more about the timeline than another benchmark will.

Tags:AnthropicJack ClarkAI safetyAGI predictionsautomated AI researchrecursive self-improvementClaudeAI researchImport AI
Oliver Senti

Oliver Senti

Senior AI Editor

Former software engineer turned tech writer, Oliver has spent the last five years tracking the AI landscape. He brings a practitioner's eye to the hype cycles and genuine innovations defining the field, helping readers separate signal from noise.

Related Articles

Stay Ahead of the AI Curve

Get the latest AI news, reviews, and deals delivered straight to your inbox. Join 100,000+ AI enthusiasts.

By subscribing, you agree to our Privacy Policy. Unsubscribe anytime.

Anthropic's Clark: 60% Odds AI Trains Successor by 2028 | aiHola