Skip to content
Home » Anthropic CEO Dario Amodei’s Interview on Dwarkesh Podcast (Transcript)

Anthropic CEO Dario Amodei’s Interview on Dwarkesh Podcast (Transcript)

Editor’s Notes: In this in-depth interview, Dwarkesh Patel sits down with Anthropic CEO Dario Amodei to explore the rapidly evolving landscape of artificial intelligence and the “big blob of compute” hypothesis. Amodei provides a rare look at the technical strategy behind Anthropic, discussing why he believes we are nearing the “end of the exponential” for AI scaling and what that means for the path to AGI. The conversation covers a wide range of topics, from the transformative impact of AI on software engineering to the internal leadership philosophy that guides one of the world’s leading AI labs. This discussion offers essential insights for anyone looking to understand the technical challenges and philosophical questions defining the current AI frontier. (Feb 13, 2026) 

TRANSCRIPT:

The Biggest Update in Three Years

DWARKESH PATEL: So we talked three years ago. I’m curious, in your view, what has been the biggest update of the last three years? What has been the biggest difference between what it felt like last three years versus now?

DARIO AMODEI: Yeah, I would say actually the underlying technology, like the exponential of the technology has gone broadly speaking, I would say about as I expected it to go. I mean there’s plus or minus a year or two here, there’s plus or minus a year or two there. I don’t know that I would have predicted the specific direction of code, but actually when I look at the exponential, it is roughly what I expected in terms of the march of the models from smart high school student to smart college student to beginning to do PhD and professional stuff. And in the case of code, reaching beyond that.

So the frontier is a little bit uneven. It’s roughly what I expected. I will tell you though what the most surprising thing has been. The most surprising thing has been the lack of public recognition of how close we are to the end of the exponential. To me it is absolutely wild that you have within the bubble and outside the bubble, but you have people talking about just the same tired old hot button political issues and around us for near the end of the exponential.

Understanding the Current Scaling Hypothesis

DWARKESH PATEL: I want to understand what that exponential looks like right now because the first question I asked you when we recorded three years ago was what’s up with scaling? Why does it work? And I have a similar question now, but I feel like it’s a more complicated question because at least from the public’s point of view, yes, three years ago there were these well known public trends where across many orders of magnitude of compute you could see how the loss improves.

And now we have RL scaling and there’s no publicly known scaling law for it. It’s not even clear what exactly the story is of is this supposed to be teaching the model skills? Is it supposed to be teaching meta learning? What is the scaling hypothesis at this point?

DARIO AMODEI: Yeah, so I have actually the same hypothesis that I had even all the way back in 2017. So in 2017 I think I talked about it last time, but I wrote a doc called “The Big Blob of Compute Hypothesis.” And it wasn’t about the scaling of language models in particular when I wrote it. GPT-1 had just come out. So that was one among many things.

Back in those days there was robotics. People tried to work on reasoning as a separate thing from language models. There was scaling of the kind of RL that happened in AlphaGo and that happened at Dota, at OpenAI and people remember StarCraft, at DeepMind, the AlphaStar. So it was written as a more general document.

And the specific thing I said was the following. Rich Sutton put out “The Bitter Lesson” a couple of years later. But the hypothesis is basically the same. So what it says is all the cleverness, all the techniques, all the “we need a new method to do something” like that doesn’t matter very much. There are only a few things that matter and I think I listed seven of them.

One is how much raw compute you have. The other is the quantity of data that you have. Then the third is the quality and distribution of data. It needs to be a broad distribution of data. The fourth is I think how long you train for. The fifth is you need an objective function that can scale to the moon. So the pre-training objective function is one such objective function.

Another objective function is the kind of RL objective function that says you have a goal, you’re going to go out and reach the goal. Within that of course there’s objective rewards like you see in math and coding and there’s more subjective rewards like you see in RL from human feedback or higher order versions of that. And then the sixth and seventh were things around normalization or conditioning, like just getting the numerical stability so that the big blob of compute flows in this laminar way instead of running into problems.

So that was the hypothesis and it’s a hypothesis I still hold. I don’t think I’ve seen very much that is not in line with that hypothesis. And so the pre-trained scaling laws were one example of what we see there. And indeed those have continued going. I think now it’s been widely reported like we feel good about pre-training, like pre-training is continuing to give us gains.

What has changed is that now we’re also seeing the same thing for RL. So we’re seeing a pre-training phase and then we’re seeing an RL phase on top of that. And with RL it’s actually just the same. Even other companies have published in some of their releases things that say look, we train the model on math contests, AIME or other things and how well the model does is log linear and how long we’ve trained it and we see that as well.