All Episodes

July 14, 2025 ‱ 90 mins

Renowned podcaster Dwarkesh Patel joins us to explore the "scaling era" of AI, characterized by rapid growth and significant compute investments. He discusses the impact of neural networks and transformers, the implications of scaling laws, and potential constraints as we approach artificial general intelligence (AGI).

Patel shares his skepticism about whether current AI models exhibit true intelligence, addresses ethical concerns around AI safety, and emphasizes the responsibilities of developers. The conversation touches on geopolitical dynamics, with major players like the U.S. and China shaping the future.

Concluding with a cautious outlook, Patel suggests a 60% chance of AGI by 2040 and highlights the importance of navigating AI complexities thoughtfully.

------
đŸ’« LIMITLESS | SUBSCRIBE & FOLLOW
https://limitless.bankless.com/
https://x.com/LimitlessFT

------
TIMESTAMPS

00:00 Start
02:15 Introduction to Dwarkesh Patel
05:45 Defining the Scaling Era
12:00 Compute
20:00 Neural Networks and Human Intelligence
28:30 Reasoning Limits of Current AI
35:40 Implications of Energy Shortages
42:10 Human-AI Relationships
51:00 AI Alignment and Moral Responsibility
01:02:15 Geopolitical Considerations
01:12:30 AI Accountability and Governance
01:20:00 The Road Ahead for AI
01:30:00 Closing Remarks

------
RESOURCES

Dwarkesh: 
https://x.com/dwarkesh_sp

The Scaling Era: An Oral History of AI, 2019–2025:
https://www.amazon.com/Scaling-Era-Oral-History-2019-2025/dp/1953953557

------
Not financial or tax advice. See our investment disclosures here:
https://www.bankless.com/disclosures⁠

Mark as Played
Transcript

Episode Transcript

Available transcripts are automatically generated. Complete accuracy is not guaranteed.
Ryan Sean Adams: Dwarkesh Patel, we are big fans. It's an honor to have you. (00:03):
undefined

Dwarkesh: Thank you so much for having me on. (00:07):
undefined

Ryan Sean Adams: Okay, so you have a book out. It's called The Scaling Era, an oral history of AI from 2019 to 2025. (00:08):
undefined

Ryan Sean Adams: These are some key dates here. This is really a story of how AI emerged. (00:16):
undefined

Ryan Sean Adams: And it seemed to have exploded on people's radar over the past five years. (00:21):
undefined

Ryan Sean Adams: And And everyone in the world, it feels like, is trying to figure out what just (00:26):
undefined

Ryan Sean Adams: happened and what is about to happen. (00:29):
undefined

Ryan Sean Adams: And I feel like for this story, we should start at the beginning, as your book does. (00:32):
undefined

Ryan Sean Adams: What is the scaling era of AI and when abouts did it start? What were the key milestones? (00:36):
undefined

Dwarkesh: So I think the undertold story about everybody's, of course, (00:42):
undefined

Dwarkesh: been hearing more and more about AI. (00:47):
undefined

Dwarkesh: The under-told story is that the big contributor to these AI models getting (00:48):
undefined

Dwarkesh: better over time has been the fact that we are throwing exponentially more compute (00:53):
undefined

Dwarkesh: into trading frontier systems every year. (00:58):
undefined

Dwarkesh: So by some estimates, we spend 4x every single year over the last decade trading (01:00):
undefined

Dwarkesh: the frontier system than the one before it. (01:04):
undefined

Dwarkesh: And that just means that we're spending hundreds of thousands of times more (01:06):
undefined

Dwarkesh: compute than the systems of the early 2010s. (01:10):
undefined

Dwarkesh: Of course, we've also had algorithmic breakthroughs in the meantime. (01:15):
undefined

Dwarkesh: 2018, we had the Transformer. (01:17):
undefined

Dwarkesh: Since then, obviously, many companies have made small improvements here and there. (01:19):
undefined

Dwarkesh: But the overwhelming fact that we're spending already hundreds of billions of (01:23):
undefined

Dwarkesh: dollars in building up the infrastructure, (01:28):
undefined

Dwarkesh: the data centers, the chips for these models, and this picture is only going (01:31):
undefined

Dwarkesh: to intensify if this exponential keeps going, (01:36):
undefined

Dwarkesh: 4x a year, over the next two years, is something that is on the minds of the (01:38):
undefined

Dwarkesh: CFOs of the big hyperscalers and the people planning the expenditures and training going forward, (01:45):
undefined

Dwarkesh: but is not as common in the conversation around where AI is headed. (01:49):
undefined

Ryan Sean Adams: So what do you feel like people should know about this? (01:54):
undefined

Ryan Sean Adams: Like what is the scaling era? There have been other eras maybe of AI or compute, (01:57):
undefined

Ryan Sean Adams: but what's special about the scaling era? (02:02):
undefined

Dwarkesh: People started noticing. Well, first of all, in 2012, there's this, (02:04):
undefined

Dwarkesh: Ilya Seskaver and others started using neural networks in order to categorize images. (02:09):
undefined

Dwarkesh: And just noticing that instead of doing something hand-coded, (02:16):
undefined

Dwarkesh: you can get a lot of juice out of just neural networks, black boxes. (02:19):
undefined

Dwarkesh: You just train them to identify what thing is like what. (02:24):
undefined

Dwarkesh: And then people started playing around these neural networks more, (02:28):
undefined

Dwarkesh: using them for different kinds of applications. (02:31):
undefined

Dwarkesh: And then the question became, we're noticing that these models get better if (02:33):
undefined

Dwarkesh: you throw more data at them and you throw more compute at them. (02:39):
undefined

Dwarkesh: How can we shove as much compute into these models as possible? (02:41):
undefined

Dwarkesh: And the solution ended up being obviously internet text. So you need an architecture (02:47):
undefined

Dwarkesh: which is amenable to the trillions of tokens that have been written over the (02:51):
undefined

Dwarkesh: last few decades and put up on the internet. (02:55):
undefined

Dwarkesh: And we had this happy coincidence of the kinds of architectures that are amenable (02:57):
undefined

Dwarkesh: to this kind of training with the GPUs that were originally made for gaming. (03:01):
undefined

Dwarkesh: We've had decades of internet text being compiled and Ilias actually called it the fossil fuel of AI. (03:05):
undefined

Dwarkesh: It's like this reservoir that we can call upon to train these minds, (03:12):
undefined

Dwarkesh: which are like, you know, they're fitting the mold of human thought because (03:17):
undefined

Dwarkesh: they're trading on trillions of tokens of human thought. (03:20):
undefined

Dwarkesh: And so then it's just been a question of making these models bigger, (03:23):
undefined

Dwarkesh: of using this data that we're getting from internet techs to further keep training them. (03:27):
undefined

Dwarkesh: And over the last year, as you know, the last six months, the new paradigm has (03:33):
undefined

Dwarkesh: been not only are we going to pre-train on all this internet text, (03:38):
undefined

Dwarkesh: we're going to see if we can have them solve math puzzles, (03:41):
undefined

Dwarkesh: coding puzzles, and through this, give them reasoning capabilities. (03:45):
undefined

Dwarkesh: The kind of thing, by the way, I mean, I have some skepticism around AGI just (03:50):
undefined

Dwarkesh: around the corner, which we'll get into. (03:55):
undefined

Dwarkesh: But just the fact that we now have machines which can like reason, (03:56):
undefined

Dwarkesh: like, you know, you can like ask a question to a machine and it'll go away for a long time. (04:00):
undefined

Dwarkesh: It'll like think about it and then like it'll come back to you with a smart answer. (04:04):
undefined

Dwarkesh: And we just sort of take it for granted. But obviously, we also know that they're (04:06):
undefined

Dwarkesh: extremely good at coding, especially. (04:10):
undefined

Dwarkesh: I don't know if you actually got a chance to play around with Cloud Code or (04:12):
undefined

Dwarkesh: Cursor or something. But it's a wild experience to design, explain at a high (04:15):
undefined

Dwarkesh: level, I want an application to does X. (04:20):
undefined

Dwarkesh: 15 minutes later, there's like 10 files of code and the application is built. (04:22):
undefined

Josh Kale: That's where we stand. (04:28):
undefined

Dwarkesh: I have takes on how much this can continue. The other important dynamic, (04:29):
undefined

Dwarkesh: I'll add my monologue here, but the other important dynamic is that if we're (04:33):
undefined

Dwarkesh: going to be living in the scaling era, you can't continue exponentials forever, (04:36):
undefined

Dwarkesh: and certainly not exponentials that are 4x a year forever. (04:41):
undefined

Dwarkesh: And so right now, we're approaching a point where within by 2028, (04:44):
undefined

Dwarkesh: at most by 2030, we will literally run out of the energy we need to keep trading (04:50):
undefined

Dwarkesh: these frontier systems, (04:57):
undefined

Dwarkesh: the capacity at the leading edge nodes, which manufacture the chips that go (04:58):
undefined

Dwarkesh: into the dyes, which go into these GPUs, even the raw fraction of GDP that will (05:02):
undefined

Dwarkesh: have to use to train frontier systems. (05:07):
undefined

Dwarkesh: So we have a couple more years left of the scaling era. And the big question (05:09):
undefined

Dwarkesh: is, will we get to AGI before then? (05:12):
undefined

Ryan Sean Adams: I mean, that's kind of a key insight of your book that like, (05:15):
undefined

Ryan Sean Adams: we're in the middle of the scaling era. (05:18):
undefined

Ryan Sean Adams: I guess we're like, you know, six years in or so. And we're not quite sure. (05:19):
undefined

Ryan Sean Adams: It's like, like the protagonist in the middle of the story, We don't know exactly (05:23):
undefined

Ryan Sean Adams: which way things are going to go. (05:26):
undefined

Ryan Sean Adams: But I want you to maybe, Dworkesh, help folks get an intuition for why scaling in this way even works. (05:28):
undefined

Ryan Sean Adams: Because I'll tell you, for me and for most people, our experience with these (05:36):
undefined

Ryan Sean Adams: revolutionary AI models probably started in 2022 with ChatGPT3 and then ChatGPT4 (05:41):
undefined

Ryan Sean Adams: and seeing all the progress, all these AI models. (05:46):
undefined

Ryan Sean Adams: And it just seems really unintuitive that if you take a certain amount of compute (05:49):
undefined

Ryan Sean Adams: and you take a certain amount of data, out pops AI, out pops intelligence. (05:55):
undefined

Ryan Sean Adams: Could you help us get an intuition for this magic? (06:01):
undefined

Ryan Sean Adams: How does the scaling law even work? Compute plus data equals intelligence? Is that really all it is? (06:05):
undefined

Dwarkesh: To be honest, I've asked so many AI researchers this exact question on my podcast. (06:12):
undefined

Dwarkesh: And I could tell you some potential theories of why it might work. (06:17):
undefined

Dwarkesh: I don't think we understand. (06:20):
undefined

Dwarkesh: You know what? I'll just say that. I don't think we understand. (06:24):
undefined

Ryan Sean Adams: We don't understand how this works. We know it works, but we don't understand (06:27):
undefined

Dwarkesh: How it works. We have evidence from actually, of all things, (06:30):
undefined

Dwarkesh: primatology of what could be going on here, or at least like why similar patterns (06:36):
undefined

Dwarkesh: in other parts of the world. (06:41):
undefined

Dwarkesh: So what I found really interesting, There's this research by this researcher, (06:42):
undefined

Dwarkesh: Susanna Herculana Huzel, (06:46):
undefined

Dwarkesh: which shows that if you look at how the number of neurons in the brain of a rat, (06:48):
undefined

Dwarkesh: different kinds of rat species increases, as the weight of their brains increase (06:56):
undefined

Dwarkesh: from species to species, there's this very sublinear pattern. (07:01):
undefined

Dwarkesh: So if their brain size doubles, the neuron count will not double between different rat species. (07:04):
undefined

Dwarkesh: And there's other animals where there's other kinds of... (07:09):
undefined

Dwarkesh: Families of species for which this is true. The two interesting exceptions to (07:14):
undefined

Dwarkesh: this rule, where there is actually a linear increase in neuron count and brain (07:18):
undefined

Dwarkesh: size, is one, certain kinds of birds. (07:22):
undefined

Dwarkesh: So, you know, birds are actually very smart, given the size of their brains, and primates. (07:26):
undefined

Dwarkesh: So the theory for what happened with humans is that we unlocked an architecture that was very scalable. (07:32):
undefined

Dwarkesh: So the way people talk about transformers being more scalable and then LSTMs, (07:38):
undefined

Dwarkesh: the thing that preceded them in 2018. (07:41):
undefined

Dwarkesh: We unlocked this architecture as it's very scalable. (07:43):
undefined

Dwarkesh: And then we were in an evolutionary niche millions of years ago, (07:46):
undefined

Dwarkesh: which rewarded marginal increases in intelligence. (07:50):
undefined

Dwarkesh: If you get slightly smarter, yes, the brain costs more energy, (07:53):
undefined

Dwarkesh: but you can save energy in terms of like not having to, you can cook, (07:56):
undefined

Dwarkesh: you can cook food so you don't have to spend much more on digestion. (07:59):
undefined

Dwarkesh: You can find a game, you can find different ways of foraging. (08:02):
undefined

Dwarkesh: Birds were not able to find this evolutionary niche, which rewarded the incremental (08:06):
undefined

Dwarkesh: increases in intelligence because if your brain gets too heavy as a bird, you're not going to fly. (08:11):
undefined

Dwarkesh: So it was this happy coincidence of these two things. Now, why is it the case (08:17):
undefined

Dwarkesh: that the fact that our brains could get bigger resulted in us becoming as smart (08:21):
undefined

Dwarkesh: as we are? We still don't know. (08:27):
undefined

Dwarkesh: And there's many different dissimilarities between AIs and humans. (08:29):
undefined

Dwarkesh: While our brains are quite big, we don't need to be trained. (08:32):
undefined

Dwarkesh: You know, a human from the age they're zero to 18 is not seeing within an order (08:35):
undefined

Dwarkesh: of magnitude of the amount of information these LLMs are trained on. (08:41):
undefined

Dwarkesh: So LLMs are extremely data inefficient. (08:44):
undefined

Dwarkesh: They need a lot more data, but the pattern of scaling, I think we see in many different places. (08:46):
undefined

Ryan Sean Adams: So is that a fair kind of analog? This analog has always made sense to me. (08:53):
undefined

Ryan Sean Adams: It's just like transformers are like neurons. (08:57):
undefined

Ryan Sean Adams: You know, AI models are sort of like the human brain. (09:00):
undefined

Ryan Sean Adams: Evolutionary pressures are like gradient descent, reward algorithms and out (09:04):
undefined

Ryan Sean Adams: pops human intelligence. We don't really understand that. (09:09):
undefined

Ryan Sean Adams: We also don't understand AI intelligence, but it's basically the same principle at work. (09:13):
undefined

Dwarkesh: I think it's a super fascinating, but also very thorny question because is gradient (09:17):
undefined

Dwarkesh: intelligence like evolution? (09:23):
undefined

Dwarkesh: Well, yes, in one sense. But also when we do gradient descent on these models, (09:24):
undefined

Dwarkesh: we start off with the weights and then we're, you know, it's like learning how (09:29):
undefined

Dwarkesh: does chemistry work, how does coding work, how does math work. (09:35):
undefined

Dwarkesh: And that's actually more similar to lifetime learning, which is to say that, (09:38):
undefined

Dwarkesh: like, by the time you're already born to the time you turn 18 or 25, (09:42):
undefined

Dwarkesh: the things you learn, and that's not evolution. (09:46):
undefined

Dwarkesh: Evolution designed the system or the brain by which you can do that learning, (09:48):
undefined

Dwarkesh: but the lifetime learning itself is not evolution. And so there's also this (09:53):
undefined

Dwarkesh: interesting question of, yeah, is training more like evolution? (09:57):
undefined

Dwarkesh: In which case, actually, we might be very far from AGI because the amount of (10:01):
undefined

Dwarkesh: compute that's been spent over the course of evolution to discover the human (10:04):
undefined

Dwarkesh: brain, you know, could be like 10 to the 40 flops. There's been estimates, you know, whatever. (10:07):
undefined

Dwarkesh: I'm sure it will bore you to discover, talk about how these estimates are derived, (10:12):
undefined

Dwarkesh: but just like how much versus is it like a single lifetime, (10:15):
undefined

Dwarkesh: like going from the age of zero to the age of 18, which is closer to, (10:20):
undefined

Dwarkesh: I think, 10 to the 24 flops, which is actually less than compute than we use (10:23):
undefined

Dwarkesh: to train frontier systems. (10:26):
undefined

Dwarkesh: All right, anyways, we'll get back to more relevant questions. (10:28):
undefined

Ryan Sean Adams: Well, here's kind of a big picture question as well. (10:33):
undefined

Ryan Sean Adams: It's like I'm constantly fascinated with the metaphysical types of discussions (10:36):
undefined

Ryan Sean Adams: that some AI researchers kind of take. (10:41):
undefined

Ryan Sean Adams: Like a lot of AI researchers will talk in terms of when they describe what they're (10:43):
undefined

Ryan Sean Adams: making, we're making God. (10:48):
undefined

Ryan Sean Adams: Like why do they say things like that? What is this talk of like making God? (10:49):
undefined

Ryan Sean Adams: What does that mean? Is it just the idea that scaling laws don't cease? (10:53):
undefined

Ryan Sean Adams: And if we can, you know, scale intelligence to AGI, then there's no reason we (10:57):
undefined

Ryan Sean Adams: can't scale far beyond that and create some sort of a godlike entity. (11:03):
undefined

Ryan Sean Adams: And essentially, that's what the quest is. We're making artificial superintelligence. (11:07):
undefined

Ryan Sean Adams: We're making a god. We're making god. (11:12):
undefined

Dwarkesh: I think people focus too much on when they, I think this God discussion focuses (11:13):
undefined

Dwarkesh: too much on the hypothetical intelligence of a single copy of an AI. (11:19):
undefined

Dwarkesh: I do believe in the notion of a super intelligence, which is not just functionally, (11:27):
undefined

Dwarkesh: which is not just like, oh, it knows a lot of things, but is actually qualitatively (11:32):
undefined

Dwarkesh: different than human society. (11:36):
undefined

Dwarkesh: But the reason is not because I think it's so powerful that any one individual (11:38):
undefined

Dwarkesh: copy of AI will be as smart, but because of the collective advantages that AIs (11:42):
undefined

Dwarkesh: will have, which have nothing to do with their raw intelligence, (11:48):
undefined

Dwarkesh: but rather the fact that these models will be digital or they already are digital, (11:51):
undefined

Dwarkesh: but eventually they'll be as smart as humans at least. (11:55):
undefined

Dwarkesh: But unlike humans, because of our biological constraints, these models can be copied. (11:58):
undefined

Dwarkesh: If there's a model that has learned a lot about a specific domain, (12:03):
undefined

Dwarkesh: you can make infinite copies of it. (12:06):
undefined

Dwarkesh: And now you have an infinite copies of Jeff Dean or Ilya Satskova or Elon Musk (12:08):
undefined

Dwarkesh: or any skilled person you can think of. (12:12):
undefined

Dwarkesh: They can be merged. So the knowledge that each copy is learning can be amalgamated (12:15):
undefined

Dwarkesh: back into the model and then back to all the copies. (12:21):
undefined

Dwarkesh: They can be distilled. They can run at superhuman speeds. (12:24):
undefined

Dwarkesh: These collective advantages, also they can communicate in latent space. (12:29):
undefined

Dwarkesh: These collective advantages. (12:32):
undefined

Ryan Sean Adams: They're immortal. I mean, you know, as an example. (12:33):
undefined

Dwarkesh: Yes, exactly. No, I mean, that's actually, tell me if I'm rabbit holing too (12:36):
undefined

Dwarkesh: much, but like one really interesting question will come about is how do we prosecute AIs? (12:40):
undefined

Dwarkesh: Because the way we prosecute humans is that we will throw you in jail if you commit a crime. (12:45):
undefined

Dwarkesh: But if there's trillions of copies or thousands of copies of an AI model, (12:50):
undefined

Dwarkesh: if a copy of an AI model, if an instance of an AI model does something bad, what do you do? (12:57):
undefined

Dwarkesh: Does the whole model have to get, and how do you even punish a model, (13:01):
undefined

Dwarkesh: right? Like, does it care about its weights being squandered? (13:04):
undefined

Dwarkesh: Yeah, there's all kinds of questions that arise because of the nature of what AIs are. (13:09):
undefined

Dwarkesh Patel: And also who is liable for that, right? (13:14):
undefined

Dwarkesh: Like, is it the toolmaker? (13:16):
undefined

Dwarkesh Patel: Is it the person using the tool? Who is responsible for these things? (13:17):
undefined

Dwarkesh Patel: There's one topic that I do want to come to here about scaling laws, (13:20):
undefined

Dwarkesh Patel: At what time did we realize that scaling laws were going to work? (13:23):
undefined

Dwarkesh Patel: Because there were a lot of theses early in the days, early 2000s about AI, (13:27):
undefined

Dwarkesh Patel: how we were going to build better models. (13:31):
undefined

Dwarkesh Patel: Eventually, we got to the transformer. But at what point did researchers and (13:33):
undefined

Dwarkesh Patel: engineers start to realize that, hey, this is the correct idea. (13:36):
undefined

Dwarkesh Patel: We should start throwing lots of money and resources towards this versus other (13:39):
undefined

Dwarkesh Patel: ideas that were just kind of theoretical research ideas, but never really took off. (13:42):
undefined

Dwarkesh Patel: We kind of saw this with GPT two to three, where there's this huge improvement. (13:45):
undefined

Dwarkesh: A lot of. (13:49):
undefined

Dwarkesh Patel: Resources went into it. Was there a specific moment in time or a specific breakthrough (13:49):
undefined

Dwarkesh Patel: that led to the start of these scaling laws? (13:53):
undefined

Dwarkesh: I think it's been a slow process of more and more people appreciating this nature (13:55):
undefined

Dwarkesh: of the overwhelming role of compute in driving forward progress. (14:00):
undefined

Dwarkesh: In 2018, I believe, Dario Amadei wrote a memo that was secret while he was at (14:05):
undefined

Dwarkesh: OpenAI. Now he's the CEO of Anthropic. (14:13):
undefined

Dwarkesh: But while he's at OpenAI, he's subsequently revealed on my podcast that he wrote (14:16):
undefined

Dwarkesh: this memo where the title of the memo was called Big Blob of Compute. (14:19):
undefined

Dwarkesh: And it says basically what you expect it to say, which is that like, (14:25):
undefined

Dwarkesh: yes, there's ways you can mess up the process of training. You have the wrong (14:29):
undefined

Dwarkesh: kinds of data or initializations. (14:31):
undefined

Dwarkesh: But fundamentally, AGI is just a big blob of compute. (14:33):
undefined

Dwarkesh: And then we've gotten over the subsequent years, there was more empirical evidence. (14:37):
undefined

Dwarkesh: So a big update, I think it was 2021. (14:41):
undefined

Dwarkesh: Correct me. Somebody definitely will correct me in the comments. (14:44):
undefined

Dwarkesh: I'm wrong. There were these, (14:46):
undefined

Dwarkesh: there's been multiple papers of these scaling laws where you can show that the (14:48):
undefined

Dwarkesh: loss of the model on the objective of predicting the next token goes down very predictably, (14:54):
undefined

Dwarkesh: almost to like multiple decimal places of correctness based on how much more (15:01):
undefined

Dwarkesh: compute you throw in these models. (15:07):
undefined

Dwarkesh: And the compute itself is a function of the amount of data you use and how big (15:08):
undefined

Dwarkesh: the model is, how many parameters it has. (15:13):
undefined

Dwarkesh: And so that was an incredibly strong evidence back in the day, (15:15):
undefined

Dwarkesh: a couple of years ago, because then you could say, well, OK, (15:19):
undefined

Dwarkesh: if it really has this incredibly low loss of predicting the next token in all (15:22):
undefined

Dwarkesh: human output, including scientific papers, including GitHub repositories. (15:28):
undefined

Dwarkesh: Then doesn't it mean it has actually had to learn coding and science and all (15:34):
undefined

Dwarkesh: these skills in order to make those predictions, which actually ended up being true. (15:40):
undefined

Dwarkesh: And it was it was something people, you know, we take it for granted now, (15:43):
undefined

Dwarkesh: but it actually even as of a year or two ago, people were really even denying that premise. (15:46):
undefined

Dwarkesh: But some people a couple of years ago just like thought about it and like, (15:50):
undefined

Dwarkesh: yeah, actually, that would mean that it's learned the skills. (15:53):
undefined

Dwarkesh: And that's crazy that we just have this strong empirical pattern that tells (15:55):
undefined

Dwarkesh: us exactly what we need to do in order to learn these skills. (15:59):
undefined

Dwarkesh Patel: And it creates this weird perception, right, where like very early on and so (16:02):
undefined

Dwarkesh Patel: to this day, it really is just a token predictor, right? (16:05):
undefined

Dwarkesh Patel: Like we're just predicting the next word in the sentence. But somewhere along (16:07):
undefined

Dwarkesh Patel: the lines, it actually creates this perception of intelligence. (16:10):
undefined

Dwarkesh Patel: So I guess we covered the early historical context. I kind of want to bring (16:14):
undefined

Dwarkesh Patel: the listeners up to today, where we are currently, where the scaling laws have (16:18):
undefined

Dwarkesh Patel: brought us in the year 2025. (16:21):
undefined

Dwarkesh Patel: So can you kind of outline where we've gotten to from early days of GPTs to (16:23):
undefined

Dwarkesh Patel: now we have GPT-4, we have Gemini Ultra, we have Club, which you mentioned earlier. (16:28):
undefined

Dwarkesh Patel: We had the breakthrough of reasoning. (16:32):
undefined

Dwarkesh Patel: So what can leading frontier models do today? (16:33):
undefined

Dwarkesh: So there's what they can do. And then there's the question of what methods seem to be working. (16:36):
undefined

Dwarkesh: I guess we can start at what they seem to be able to do. They've shown to be (16:41):
undefined

Dwarkesh: remarkably useful at coding and not just at answering direct questions about (16:46):
undefined

Dwarkesh: how does this line of code work or something. (16:51):
undefined

Dwarkesh: But genuinely just autonomously working for 30 minutes or an hour, (16:53):
undefined

Dwarkesh: doing the task, it would take a front-end developer a whole day to do. (16:57):
undefined

Dwarkesh: And you can just ask them at a high level, do this kind of thing, (17:01):
undefined

Dwarkesh: and they can go ahead and do it. (17:04):
undefined

Dwarkesh: Obviously, if you've played around with it, you know that they're extremely (17:05):
undefined

Dwarkesh: useful assistants in terms of research, in terms of even therapists, (17:07):
undefined

Dwarkesh: whatever other use cases. (17:12):
undefined

Dwarkesh: On the question of what training methods seem to be working, (17:13):
undefined

Dwarkesh: we do seem to be getting evidence that pre-training is plateauing, (17:16):
undefined

Dwarkesh: which is to say that we had GPT 4.5, which was just following this old mold (17:19):
undefined

Dwarkesh: of make the model bigger, (17:25):
undefined

Dwarkesh: but it's fundamentally doing the same thing of next token prediction. (17:27):
undefined

Dwarkesh: And apparently it didn't pass muster. The OpenAI had to deprecate it because (17:31):
undefined

Dwarkesh: there's this dynamic where the bigger the model is, the more it costs not only (17:36):
undefined

Dwarkesh: to train, but also to serve, right? (17:39):
undefined

Dwarkesh: Because every time you serve a user, you're having to run the whole model, (17:41):
undefined

Dwarkesh: which is going, so, but that doesn't be working is RL, which is this process (17:44):
undefined

Dwarkesh: of, not just training them on existing tokens on the internet, (17:49):
undefined

Dwarkesh: but having the model itself try to answer math and coding problems. (17:52):
undefined

Dwarkesh: And finally, we got to the point where the model is smart enough to get it right (17:55):
undefined

Dwarkesh: some of the time, and so you can give it some reward, and then it can saturate (17:57):
undefined

Dwarkesh: these tough reasoning problems. (18:01):
undefined

Dwarkesh Patel: And then what was the breakthrough with reasoning for the people who aren't familiar? (18:04):
undefined

Dwarkesh Patel: What made reasoning so special that we hadn't discovered before? (18:08):
undefined

Dwarkesh Patel: And what did that kind of unlock for models that we use today? (18:11):
undefined

Dwarkesh: I'm honestly not sure. I mean, GBD-4 came out a little over two years ago, (18:14):
undefined

Dwarkesh: and then it was after two years after GPT-4 came out that O-1 came out which (18:19):
undefined

Dwarkesh: was the original reasoning breakthrough I think last November and, (18:23):
undefined

Dwarkesh: And subsequently, a couple of months later, DeepSeq showed in their R1 paper. (18:28):
undefined

Dwarkesh: So DeepSeq open source their research and they explained exactly how their algorithm worked. (18:33):
undefined

Dwarkesh: And it wasn't that complicated. It was just like what you would expect, (18:37):
undefined

Dwarkesh: which is get some math problems, (18:41):
undefined

Dwarkesh: give for some initial problems, tell the model exactly what the reasoning trace (18:44):
undefined

Dwarkesh: looks like, how you solve it, just like write it out and then have the model (18:48):
undefined

Dwarkesh: like try to do it raw on the remaining problems. (18:51):
undefined

Dwarkesh: Now, I know it sounds incredibly arrogant to say, well, it wasn't that complicated. (18:54):
undefined

Dwarkesh: Why did it take you years? (18:57):
undefined

Dwarkesh: I think there's an interesting insight there of even things which you think (18:58):
undefined

Dwarkesh: will be simple in terms of high level description of how to solve the problem (19:02):
undefined

Dwarkesh: end up taking longer in terms of haggling out the remaining engineering hurdles (19:05):
undefined

Dwarkesh: than you might naively assume. (19:10):
undefined

Dwarkesh: And that should update us on how long it will take us to go through the remaining (19:12):
undefined

Dwarkesh: bottlenecks on the path to AGI. (19:17):
undefined

Dwarkesh: Maybe that will be tougher than people imagine, especially the people who think (19:19):
undefined

Dwarkesh: we're only two to three years away. (19:22):
undefined

Dwarkesh: But all this to say, yeah, I'm not sure why it took so long after GPT-4 to get (19:24):
undefined

Dwarkesh: a model trained on a similar level of capabilities that could then do reasoning. (19:27):
undefined

Dwarkesh Patel: And in terms of those abilities, the first answer you had to what can it do was coding. (19:31):
undefined

Dwarkesh Patel: And I hear that a lot of the time when I talk to a lot of people is that coding (19:36):
undefined

Dwarkesh Patel: seems to be a really strong suit and a really huge unlock to using these models. (19:40):
undefined

Dwarkesh Patel: And I'm curious, why coding over general intelligence? (19:44):
undefined

Dwarkesh Patel: Is it because it's placed in a more confined box of parameters? (19:48):
undefined

Dwarkesh Patel: I know in the early days, we had the AlphaGo and And we had the AIs playing (19:51):
undefined

Dwarkesh Patel: chess and they exceed, they perform so well because they were kind of contained (19:54):
undefined

Dwarkesh Patel: within this box of parameters that was a little less open-ended than general intelligence. (19:58):
undefined

Dwarkesh Patel: Is that the reason why coding is kind of at the frontier right now of the ability of these models? (20:01):
undefined

Dwarkesh: There's two different hypotheses. One is based around this idea called Moravac's paradox. (20:06):
undefined

Dwarkesh: And this was an idea, by the way, one super interesting figure, (20:13):
undefined

Dwarkesh: actually, I should have mentioned him earlier. (20:16):
undefined

Dwarkesh: One super interesting figure in the history of scaling is Hans Moravac, (20:18):
undefined

Dwarkesh: who I think in the 90s predicts that 2028 will be the year that we will get to AGI. (20:22):
undefined

Dwarkesh: And the way he predicts this, which is like, you know, we'll see what happens, (20:29):
undefined

Dwarkesh: but like not that far off the money as far as I'm concerned. (20:32):
undefined

Dwarkesh: The way he predicts this is he just looks at the growth in computing power year (20:35):
undefined

Dwarkesh: over year and then looks at how much compute he estimated the human brain to be to require. (20:41):
undefined

Dwarkesh: And just like, OK, we'll have computers as powerful as the human brain by 2028. (20:47):
undefined

Dwarkesh: Which is like at once a deceptively simple argument, but also ended up being (20:51):
undefined

Dwarkesh: incredibly accurate and like worked, right? (20:57):
undefined

Dwarkesh: I might add a fact drive it was 2028, but it was within that, (21:01):
undefined

Dwarkesh: like within something you would consider a reasonable guess, given what we know now. (21:03):
undefined

Dwarkesh: Sorry, anyway, so the Morrowind's paradox is this idea that computers seemed (21:07):
undefined

Dwarkesh: in AI get better first at the skills which humans are the worst at. (21:12):
undefined

Dwarkesh: Or at least there's a huge variation in the human repertoire. (21:19):
undefined

Dwarkesh: So we think of coding as incredibly hard, right? We think this is like the top (21:22):
undefined

Dwarkesh: 1% of people will be excellent coders. (21:26):
undefined

Dwarkesh: We also think of reasoning as very hard, right? So if you like read Aristotle, (21:29):
undefined

Dwarkesh: he says, the thing which makes humans special, which distinguishes us from animals is reasoning. (21:32):
undefined

Dwarkesh: And these models aren't that useful yet at almost anything. The one thing they can do is reasoning. (21:38):
undefined

Dwarkesh: So how do we explain this pattern? And Moravec's answer is that evolution has (21:45):
undefined

Dwarkesh: spent billions of years optimizing us to do things we take for granted. (21:51):
undefined

Dwarkesh: Move around this room, right? I can pick up this can of Coke, (21:56):
undefined

Dwarkesh: move it around, drink from it. (21:58):
undefined

Dwarkesh: And that we can't even get robots to do at all yet. (22:00):
undefined

Dwarkesh: And in fact, it's so ingrained in us by evolution that there's no human, or. (22:04):
undefined

Ryan Sean Adams: At least humans who don't have (22:08):
undefined

Dwarkesh: Disabilities will all be able to do this. And so we just take it for granted (22:10):
undefined

Dwarkesh: that this is an easy thing to do. (22:13):
undefined

Dwarkesh: But in fact, it's evidence of how long evolution has spent getting humans up to this point. (22:15):
undefined

Dwarkesh: Whereas reasoning, logic, all of these skills have only been optimized by evolution (22:19):
undefined

Dwarkesh: over the course of the last few million years. (22:27):
undefined

Dwarkesh: So there's been a thousand fold less evolutionary pressure towards coding than (22:30):
undefined

Dwarkesh: towards just basic locomotion. (22:35):
undefined

Dwarkesh: And this has actually been very accurate in predicting what kinds of progress (22:38):
undefined

Dwarkesh: we see even before we got deep learning, right? (22:42):
undefined

Dwarkesh: Like in the 40s when we got our first computers, the first thing that we could (22:43):
undefined

Dwarkesh: use them to do is long calculations for ballistic trajectories at the time for World War II. (22:47):
undefined

Dwarkesh: Humans suck at long calculations by hand. (22:52):
undefined

Dwarkesh: And anyways, so that's the explanation for coding, which seems hard for humans, (22:56):
undefined

Dwarkesh: is the first thing that went to AIs. (23:00):
undefined

Dwarkesh: Now, there's another theory, which is that this is actually totally wrong. (23:02):
undefined

Dwarkesh: It has nothing to do with the seeming paradox of how long evolution has optimized (23:05):
undefined

Dwarkesh: us for, and everything to do with the availability of data. (23:10):
undefined

Dwarkesh: So we have GitHub, this repository of all of human code, at least all open source (23:15):
undefined

Dwarkesh: code written in all these different languages, trillions and trillions of tokens. (23:23):
undefined

Dwarkesh: We don't have an analogous thing for robotics. We don't have this pre-training (23:26):
undefined

Dwarkesh: corpus. And that explains why code has made so much more progress than robotics. (23:29):
undefined

Ryan Sean Adams: That's fascinating because if there's one thing that I could list that we'd (23:34):
undefined

Ryan Sean Adams: want AI to be good at, probably coding software is number one on that list. (23:38):
undefined

Ryan Sean Adams: Because if you have a Turing complete intelligence that can create Turing complete (23:44):
undefined

Ryan Sean Adams: software, is there anything you can't create once you have that? (23:49):
undefined

Ryan Sean Adams: Also, like the idea of Morvac's paradox, I guess that sort of implies a certain (23:52):
undefined

Ryan Sean Adams: complementarianism with humanity. (23:58):
undefined

Ryan Sean Adams: So if robots can do things that robots can do really well and can't do the things (24:01):
undefined

Ryan Sean Adams: humans can do well, well, perhaps there's a place for us in this world. (24:06):
undefined

Ryan Sean Adams: And that's fantastic news. It also maybe implies that humans have kind of scratched (24:09):
undefined

Ryan Sean Adams: the surface on reasoning potential. (24:14):
undefined

Ryan Sean Adams: I mean, if we've only had a couple of million years of evolution and we haven't (24:17):
undefined

Ryan Sean Adams: had the data set to actually get really good at reasoning, it seems like there'd (24:21):
undefined

Ryan Sean Adams: be a massive amount of upside, unexplored territory, (24:25):
undefined

Ryan Sean Adams: like so much more intelligence that nature could actually (24:29):
undefined

Ryan Sean Adams: contain inside of reasoning. (24:33):
undefined

Ryan Sean Adams: I mean, are these some of the implications of these ideas? (24:35):
undefined

Dwarkesh: Yeah, I know. I mean, that's a great insight. Another really interesting insight (24:38):
undefined

Dwarkesh: is that the more variation there (24:41):
undefined

Dwarkesh: is in a skill in humans, the better and faster that AIs will get at it. (24:44):
undefined

Dwarkesh: Because coding is the kind of thing where 1% of humans are really good at it. (24:50):
undefined

Dwarkesh: The rest of us will, if we try to learn it, we'd be okay at it or something, right? (24:55):
undefined

Dwarkesh: And because evolutionists spend so little time optimizing us, (25:00):
undefined

Dwarkesh: there's this room for variation where the optimization hasn't happened uniformly (25:03):
undefined

Dwarkesh: or it hasn't been valuable enough to saturate the human gene pool for this skill. (25:07):
undefined

Dwarkesh: I think you made an earlier point that I thought was really interesting I wanted (25:14):
undefined

Dwarkesh: to address. Can you remind me of the first thing you said? Is it the complementarianism? Yes. (25:16):
undefined

Dwarkesh: So you can take it as a positive future. You can take it as a negative future (25:23):
undefined

Dwarkesh: in the sense that, well, what is the complementary skills we're providing? (25:27):
undefined

Dwarkesh: We're good meat robots. (25:30):
undefined

Ryan Sean Adams: Yeah, the low skilled labor of the situation. (25:33):
undefined

Dwarkesh: We can do all the thinking and planning. One dark future, (25:35):
undefined

Dwarkesh: one dark vision of the future is we'll get those meta glasses (25:39):
undefined

Dwarkesh: and the AI speaking into our ear and it'll tell us to go put this brick over (25:44):
undefined

Dwarkesh: there so that the next data center couldn't be built because the AI's got the (25:50):
undefined

Dwarkesh: plan for everything. It's got the better design for the ship and everything. (25:53):
undefined

Dwarkesh: You just need to move things around for it. And that's what human labor looks (25:55):
undefined

Dwarkesh: like until robotics is solved. (25:58):
undefined

Dwarkesh: So yeah, it depends on how you... On the other hand, you'll get paid a lot because (26:01):
undefined

Dwarkesh: it's worth a lot to move those bricks. We're building AGI here. (26:04):
undefined

Dwarkesh: But yeah, it depends on how you come out of that question. (26:08):
undefined

Ryan Sean Adams: Well, there seems to be something to that idea, going back to the idea of the (26:09):
undefined

Ryan Sean Adams: massive amount of human variation. (26:12):
undefined

Ryan Sean Adams: I mean, we have just in the past month or so, we have news of meta hiring AI (26:14):
undefined

Ryan Sean Adams: researchers for $100 million signing bonuses, okay? (26:18):
undefined

Ryan Sean Adams: What does the average software engineer make versus what does an AI researcher (26:22):
undefined

Ryan Sean Adams: make at kind of the top of the market, right? (26:27):
undefined

Ryan Sean Adams: Which has got to imply, obviously there's some things going on with demand and (26:29):
undefined

Ryan Sean Adams: supply, but also that it does also seem to imply that there's massive variation (26:33):
undefined

Ryan Sean Adams: in the quality of a software engineer. (26:38):
undefined

Ryan Sean Adams: And if AIs can get to that quality, well, what does that unlock? (26:40):
undefined

Ryan Sean Adams: Yeah. So, okay. Yeah. So I guess we have like coding down right now. (26:44):
undefined

Ryan Sean Adams: Like another question though is like, what can't AIs do today? (26:48):
undefined

Ryan Sean Adams: And how would you characterize that? Like what are the things they just don't do well? (26:53):
undefined

Dwarkesh: So I've been interviewing people on my podcast who have very different timelines (26:57):
undefined

Dwarkesh: from a role to get to AGI. I have had people on who think it's two years away (27:02):
undefined

Dwarkesh: and some who think it's 20 years away. (27:05):
undefined

Dwarkesh: And the experience of building AI tools for myself actually has been the most (27:08):
undefined

Dwarkesh: insight driving or maybe research I've done on the question of when AI is coming. (27:13):
undefined

Ryan Sean Adams: More than the guest interviews. (27:18):
undefined

Dwarkesh: Yeah, because you just, I have had, I've probably spent on the order of a hundred (27:20):
undefined

Dwarkesh: hours trying to build these little tools. The kinds I'm sure you've also tried (27:25):
undefined

Dwarkesh: to build of like, rewrite auto-generated transcripts for me to make them sound, (27:28):
undefined

Dwarkesh: the rewritten the way a human would write them. (27:32):
undefined

Dwarkesh: Find clips for me to tweet out, write essays with me, co-write them passage (27:35):
undefined

Dwarkesh: by passage, these kinds of things. (27:39):
undefined

Dwarkesh: And what I found is that it's actually very hard to get human-like labor out (27:41):
undefined

Dwarkesh: of these models, even for tasks like these, which should be death center in (27:45):
undefined

Dwarkesh: the repertoire of these models, right? (27:49):
undefined

Dwarkesh: They're short horizon, they're language in, language out. (27:50):
undefined

Dwarkesh: They're not contingent on understanding some thing I said like a month ago. (27:53):
undefined

Dwarkesh: This is just like, this is the task. (27:58):
undefined

Dwarkesh: And I was thinking about why is it the case that I still haven't been able to (28:00):
undefined

Dwarkesh: automate these basic language tasks? Why do I still have a human work on these things? (28:04):
undefined

Dwarkesh: And I think the key reason that you can't automate even these simple tasks is (28:09):
undefined

Dwarkesh: because the models currently lack the ability to do on the job training. (28:15):
undefined

Dwarkesh: So if you hire a human for the first six months, for the first three months, (28:21):
undefined

Dwarkesh: they're not going to be that useful, even if they're very smart, (28:24):
undefined

Dwarkesh: because they haven't built up the context, they haven't practiced the skills, (28:26):
undefined

Dwarkesh: they don't understand how the business works. (28:29):
undefined

Dwarkesh: What makes humans valuable is not that mainly the raw intellect obviously matters, (28:31):
undefined

Dwarkesh: but it's not mainly that. (28:35):
undefined

Dwarkesh: It's their ability to interrogate their own failures in this really dynamic, (28:36):
undefined

Dwarkesh: organic way to pick up small efficiencies and improvements as they practice (28:40):
undefined

Dwarkesh: the task and to build up this context as they work within a domain. (28:45):
undefined

Dwarkesh: And so sometimes people wonder, look, if you look at the revenue of OpenAI, (28:50):
undefined

Dwarkesh: the annual recurring revenue, it's on the order of $10 billion. (28:54):
undefined

Dwarkesh: Kohl's makes more money than that. McDonald's makes more money than that, right? (28:57):
undefined

Dwarkesh: So why is it that if they've got AGI, they're, you know, like Fortune 500 isn't (29:01):
undefined

Dwarkesh: reorganizing their workflows to, you know, use open AI models at every layer of the stack? (29:07):
undefined

Dwarkesh: My answer, sometimes people say, well, it's because people are too stodgy. (29:12):
undefined

Dwarkesh: The management of these companies is like not moving fast enough on AI. (29:15):
undefined

Dwarkesh: That could be part of it. I think mostly it's not that. (29:18):
undefined

Dwarkesh: I think mostly it genuinely is very hard to get human-like labor out of these (29:20):
undefined

Dwarkesh: models because you can't. (29:23):
undefined

Dwarkesh: So you're stuck with the capabilities you get out of the model out of the box. (29:26):
undefined

Dwarkesh: So they might be five out of 10 at rewriting the transcript for you. (29:30):
undefined

Dwarkesh: But if you don't like how it turned out, if you have feedback for it, (29:33):
undefined

Dwarkesh: if you want to keep teaching it over time, once the session ends, (29:36):
undefined

Dwarkesh: the model, like everything it knows about you has gone away. (29:41):
undefined

Dwarkesh: You got to restart again. It's like working with an amnesiac employee. (29:44):
undefined

Dwarkesh: You got to restart again. (29:47):
undefined

Ryan Sean Adams: Every day is the first day of employment, basically. (29:49):
undefined

Dwarkesh: Yeah, exactly. It's a groundhog day for them every day or every couple of hours, in fact. (29:52):
undefined

Dwarkesh: And that makes it very hard for them to be that useful as an employee, (29:56):
undefined

Dwarkesh: right? They're not really an employee at that point. (29:59):
undefined

Dwarkesh: This, I think, not only is a key bottleneck to the value of these models, (30:02):
undefined

Dwarkesh: because human labor is worth a lot, right? (30:06):
undefined

Dwarkesh: Like $60 trillion in the world is paid to wages every year. (30:09):
undefined

Dwarkesh: If these model companies are making on the order of $10 billion a year, that's a big way to AGI. (30:12):
undefined

Dwarkesh: And what explains that gap? What are the bottlenecks? I think a big one is this (30:18):
undefined

Dwarkesh: continual learning thing. (30:21):
undefined

Dwarkesh: And I don't see an easy way that that just gets solved within these models. (30:23):
undefined

Dwarkesh: There's no like, with reasoning, you could say, oh, it's like train it on math (30:26):
undefined

Dwarkesh: and code problems, and then I'll get the reasoning. And that worked. (30:29):
undefined

Dwarkesh: I don't think there's something super obvious there for how do you get this (30:31):
undefined

Dwarkesh: online learning, this on-the-job training working for these models. (30:35):
undefined

Ryan Sean Adams: Okay, can we talk about that, go a little bit deeper on that concept? (30:38):
undefined

Ryan Sean Adams: So this is basically one of the concepts you wrote in your recent post. (30:41):
undefined

Ryan Sean Adams: AI is not right around the corner. Even though you're an AI optimist, (30:44):
undefined

Ryan Sean Adams: I would say, and overall an AI accelerationist, you You were saying it's not (30:48):
undefined

Ryan Sean Adams: right around the corner. (30:53):
undefined

Ryan Sean Adams: You're saying the ability to replace human labor is a ways out. (30:54):
undefined

Ryan Sean Adams: Not forever out, but I think you said somewhere around 2032, (30:58):
undefined

Ryan Sean Adams: if you had to guess on when the estimate was. (31:01):
undefined

Ryan Sean Adams: And the reason you gave is because AIs can't learn on the job, (31:05):
undefined

Ryan Sean Adams: but it's not clear to me why they can't. (31:09):
undefined

Ryan Sean Adams: Is it just because the context window isn't large enough? (31:10):
undefined

Ryan Sean Adams: Is it just because they can't input all of the different data sets and data (31:14):
undefined

Ryan Sean Adams: points that humans can? Is it because they don't have stateful memory the way a human employee? (31:19):
undefined

Ryan Sean Adams: Because if it's these things, all of these do seem like solvable problems. (31:25):
undefined

Ryan Sean Adams: And maybe that's what you're saying. They are solvable problems. (31:28):
undefined

Ryan Sean Adams: They're just a little bit longer than some people think they are. (31:30):
undefined

Dwarkesh: I think it's like in some deep sense a solvable problem because eventually we will build AGI. (31:35):
undefined

Dwarkesh: And to build AGI, we will have had to solve the problem. (31:40):
undefined

Dwarkesh: My point is that the obvious solutions you might imagine, for example, (31:43):
undefined

Dwarkesh: expanding the context window or having this (31:46):
undefined

Dwarkesh: like external memory using systems like rag these (31:49):
undefined

Dwarkesh: are basically techniques we already have to it's called retrieval augmented (31:54):
undefined

Dwarkesh: generate anyways these kinds of retrieval augmented generation i (31:57):
undefined

Dwarkesh: don't think these will suffice and just to put a finer point first of all like (32:00):
undefined

Dwarkesh: what is the problem the problem is exactly as you say that within the context (32:04):
undefined

Dwarkesh: window these models actually can learn on the job right so if you talk to it (32:09):
undefined

Dwarkesh: for long enough it will get much better at understanding your needs and what your exact problem is. (32:13):
undefined

Dwarkesh: If you're using it for research for your podcast, it will get a sense of like, (32:17):
undefined

Dwarkesh: oh, they're actually especially curious about these kinds of questions. Let me focus on that. (32:21):
undefined

Dwarkesh: It's actually very human-like in context, right? The speed at which it learns, (32:25):
undefined

Dwarkesh: the task of knowledge it picks out. (32:28):
undefined

Dwarkesh: The problem, of course, is the context length for even the best models only (32:30):
undefined

Dwarkesh: last a million or two million tokens. (32:33):
undefined

Dwarkesh: That's at most like an hour of conversation. (32:36):
undefined

Dwarkesh: Now, then you might say, okay, well, why can't we just solve that by expanding (32:39):
undefined

Dwarkesh: the context window, right? So context window has been expanding for the last (32:42):
undefined

Dwarkesh: few years. Why can't we just continue that? (32:44):
undefined

Ryan Sean Adams: Yeah, like a billion token context window, something like this. (32:47):
undefined

Dwarkesh: So 2018 is when the transformer came out and the transformer has the attention mechanism. (32:50):
undefined

Dwarkesh: The attention mechanism is inherently quadratic with the nature of the length (32:55):
undefined

Dwarkesh: of the sequence, which is to say that if you go from if you double go from 1 (33:00):
undefined

Dwarkesh: million tokens to 2 million tokens, (33:05):
undefined

Dwarkesh: it actually costs four times as much compute to process that 2 millionth token. (33:07):
undefined

Dwarkesh: It's not just 2 to as much compute. so it gets super linearly more expensive (33:12):
undefined

Dwarkesh: as you increase the context length and for the last, (33:18):
undefined

Dwarkesh: seven years people have been trying to get around this inherent quadratic nature (33:23):
undefined

Dwarkesh: of attention of course we don't know secretly what the labs are working on but we have frontier, (33:26):
undefined

Dwarkesh: companies like deep seek which have open source their research and (33:32):
undefined

Dwarkesh: we can just see how their algorithms work and they found (33:35):
undefined

Dwarkesh: these constant time modifiers to attention which is (33:38):
undefined

Dwarkesh: to say that they there's like a it'll still (33:41):
undefined

Dwarkesh: be quadratic but it'll be like one half times (33:44):
undefined

Dwarkesh: quadratic but the inherent like super linearness has not (33:47):
undefined

Dwarkesh: gone away and because of that yeah you might be able to increase it from 1 million (33:49):
undefined

Dwarkesh: tokens to 2 million tokens by finding another hack like uh make sure experts (33:53):
undefined

Dwarkesh: just run such things latent attention is another such technique but or kbcash (33:57):
undefined

Dwarkesh: right there's many other things that have been discovered but people have not (34:01):
undefined

Dwarkesh: discovered okay how do you get around the fact that if you went to a billion, (34:05):
undefined

Dwarkesh: it would be a billion squared as expensive in terms of compute to process that token. (34:09):
undefined

Dwarkesh: And so I don't think you'll just get it by increasing the length of the context window, basically. (34:14):
undefined

Ryan Sean Adams: That's fascinating. Yeah, I didn't realize that. Okay, so the other reason in (34:19):
undefined

Ryan Sean Adams: your post that AI is not right around the corner is because it can't do your taxes. (34:23):
undefined

Ryan Sean Adams: And Dwarkesh, I feel your pain, man. Taxes are just like quite a pain in the ass. (34:27):
undefined

Ryan Sean Adams: I think you were talking about this from the context of like computer vision, (34:33):
undefined

Ryan Sean Adams: computer use, that kind of thing. (34:36):
undefined

Ryan Sean Adams: So, I mean, I've seen demos. I've seen some pretty interesting computer vision (34:38):
undefined

Ryan Sean Adams: sort of demos that seem to be right around the corner. (34:42):
undefined

Ryan Sean Adams: But what's the limiter on computer use for an AI? (34:46):
undefined

Dwarkesh: There was an interesting blog post by this company called Mechanize where they (34:49):
undefined

Dwarkesh: were explaining why this is such a big problem. And I love the way they phrased it, which is that, (34:54):
undefined

Dwarkesh: Imagine if you had to train a model in 1980, a large language model in 1980, (34:58):
undefined

Dwarkesh: and you could use all the compute you wanted in 1980 somehow, (35:04):
undefined

Dwarkesh: but you didn't have, you were only stuck with the data that was available in (35:08):
undefined

Dwarkesh: the 1980s, of course, before the internet became a widespread phenomenon. (35:14):
undefined

Dwarkesh: You couldn't train a modern LLM, even with all the computer in the world, (35:17):
undefined

Dwarkesh: because the data wasn't available. (35:20):
undefined

Dwarkesh: And we're in a similar position with respect to computer use, (35:22):
undefined

Dwarkesh: because there's not this corpus of collected videos, people using computers (35:26):
undefined

Dwarkesh: to do different things, to access different applications and do white collar work. (35:31):
undefined

Dwarkesh: Because of that, I think the big challenge has been accumulating this kind of data. off. (35:37):
undefined

Ryan Sean Adams: And to be clear, when I was saying the use case of like, do my taxes, (35:43):
undefined

Ryan Sean Adams: you're effectively talking about an AI having the ability to just like, (35:47):
undefined

Ryan Sean Adams: you know, navigate the files around your computer, (35:51):
undefined

Ryan Sean Adams: you know, log in to various websites to download your pay stubs or whatever, (35:54):
undefined

Ryan Sean Adams: and then to go to like TurboTax or something and like input it all into some (35:58):
undefined

Ryan Sean Adams: software and file it, right? (36:02):
undefined

Ryan Sean Adams: Just on voice command or something like that. That's basically doing my taxes. (36:04):
undefined

Dwarkesh: It should be capable of navigating UIs that it's less familiar with or that (36:08):
undefined

Dwarkesh: come about organically within the context of trying to solve a problem. (36:13):
undefined

Dwarkesh: So for example, I might have business deductions. (36:17):
undefined

Dwarkesh: It sees on my bank statement that I've spent $1,000 on Amazon. (36:20):
undefined

Dwarkesh: It goes logs in my Amazon. (36:24):
undefined

Dwarkesh: It sees like, oh, he bought a camera. So I think that's probably a business (36:25):
undefined

Dwarkesh: expense for his podcast. (36:29):
undefined

Dwarkesh: He bought an Airbnb over a weekend in the cabins of whatever, (36:30):
undefined

Dwarkesh: in the woods of whatever. That probably wasn't a business expense. (36:35):
undefined

Dwarkesh: Although maybe, maybe it's, if it's a sort of like a gray, if it's willing to (36:38):
undefined

Dwarkesh: go in the gray area, maybe I'll talk to you. Yeah, yeah, yeah. (36:42):
undefined

Ryan Sean Adams: Do the gray area stuff. (36:45):
undefined

Dwarkesh: I was, I was researching. (36:46):
undefined

Dwarkesh: But anyway, so that, including all of that, including emailing people for invoices, (36:50):
undefined

Dwarkesh: and haggling with them, it would be like a sort of week long task to do my taxes, right? (36:55):
undefined

Dwarkesh: You'd have to, there's a lot of work involved. That's not just like do this (37:01):
undefined

Dwarkesh: skill, this skill, this skill, but rather of having a sort of like plan of action (37:04):
undefined

Dwarkesh: and then breaking tasks apart, dealing with new information, (37:08):
undefined

Dwarkesh: new emails, new messages, consulting with me about questions, et cetera. (37:11):
undefined

Ryan Sean Adams: Yeah, I mean, to be clear on this use case too, even though your post is titled (37:16):
undefined

Ryan Sean Adams: like, you know, AI is not right around the corner, you still think this ability (37:18):
undefined

Ryan Sean Adams: to file your taxes, that's like a 2028 thing, right? (37:22):
undefined

Ryan Sean Adams: I mean, this is maybe not next year, but it's in a few years. (37:27):
undefined

Dwarkesh: Right, which is, I think that was sort of, people maybe write too much in The (37:31):
undefined

Dwarkesh: Decital and then didn't read through the arguments. (37:35):
undefined

Ryan Sean Adams: I mean, that never happens on the internet. Wow. (37:37):
undefined

Dwarkesh: First time. (37:40):
undefined

Dwarkesh: No, I think like I'm arguing against people who are like, you know, this will happen. (37:42):
undefined

Dwarkesh: AGI is like two years away. I do think the wider world, the markets, (37:47):
undefined

Dwarkesh: public perception, even people who are somewhat attending to AI, (37:53):
undefined

Dwarkesh: but aren't in this specific milieu that I'm talking to, are way underpricing AGI. (37:57):
undefined

Dwarkesh: One reason, one thing I think they're underestimating is not only will we have (38:03):
undefined

Dwarkesh: millions of extra laborers, millions of extra workers, (38:09):
undefined

Dwarkesh: potentially billions within the course of the next decade, because then we will (38:12):
undefined

Dwarkesh: have a potentially, I think like likely we will have AGI within the next decade. (38:16):
undefined

Dwarkesh: But they'll have these advantages that human workers don't have, (38:20):
undefined

Dwarkesh: which is that, okay, a single model company, so suppose we solve continual learning, right? (38:23):
undefined

Dwarkesh: So there, and we saw computer use. So as far as white collar work goes, (38:27):
undefined

Dwarkesh: that might fundamentally it would be solved. (38:31):
undefined

Dwarkesh: You can have AIs which can use not just they're not just like a text box where (38:32):
undefined

Dwarkesh: you put into you ask questions in a chatbot and you get some response out. (38:36):
undefined

Dwarkesh: It's not that useful to just have a very smart chatbot. You need it to be able (38:39):
undefined

Dwarkesh: to actually do real work and use real applications. (38:42):
undefined

Dwarkesh: Suppose you have that solved because it acts like an employee. (38:45):
undefined

Dwarkesh: It's got continual learning. It's got computer use. (38:48):
undefined

Dwarkesh: But it has another advantage that humans don't have, which is that copies of (38:49):
undefined

Dwarkesh: this model are going being deployed all through the economy and it's doing on the job training. (38:53):
undefined

Dwarkesh: So copies are learning how to be an accountant, how to be a lawyer, (38:57):
undefined

Dwarkesh: how to be a coder, except because it's an AI and it's digital, (39:00):
undefined

Dwarkesh: the model itself can amalgamate all this on-the-job training from all these copies. (39:04):
undefined

Dwarkesh: So what does that mean? Well, it means that even if there's no more software (39:10):
undefined

Dwarkesh: progress after that point, which is to say that no more algorithms are discovered, (39:13):
undefined

Dwarkesh: there's not a transformer plus plus that's discovered. (39:17):
undefined

Dwarkesh: Just from the fact that this model is learning every single skill in the economy, (39:20):
undefined

Dwarkesh: at least for white-collar work, you might just, based on that alone, (39:25):
undefined

Dwarkesh: have something that looks like an intelligence explosion. (39:30):
undefined

Dwarkesh: It would just be a broadly deployed intelligence explosion, but it would functionally (39:31):
undefined

Dwarkesh: become super intelligent just from having human-level capability of learning on the job. (39:35):
undefined

Dwarkesh Patel: Yeah, and it creates this mesh network of intelligence that's shared among everyone. (39:41):
undefined

Dwarkesh Patel: That's a really fascinating thing. So we're going to get there. (39:45):
undefined

Dwarkesh Patel: We're going to get to AGI. it's going to be incredibly smart. (39:48):
undefined

Dwarkesh Patel: But what we've shared recently is just kind of this mixed bag where currently (39:51):
undefined

Dwarkesh Patel: today, it's pretty good at some things, but also not that great at others. (39:54):
undefined

Dwarkesh Patel: We're hiring humans to do jobs that we think AI should do, but it probably doesn't. (39:57):
undefined

Dwarkesh Patel: So the question I have for you is, is AI really that smart? Or is it just good (40:00):
undefined

Dwarkesh Patel: at kind of acing these particular benchmarks that we measure against? (40:04):
undefined

Dwarkesh Patel: Apple, I mean, famously recently, they had their paper, The Illusion of Thinking, (40:08):
undefined

Dwarkesh Patel: where it was kind of like, hey, AI is like pretty good up to a point, (40:11):
undefined

Dwarkesh Patel: but at a certain point, it just falls apart. (40:14):
undefined

Dwarkesh Patel: And the inference is like, maybe it's not intelligence, maybe it's just good (40:16):
undefined

Dwarkesh Patel: at guessing. So I guess the question is, is AI really that smart? (40:21):
undefined

Dwarkesh: It depends on who I'm talking to. I think some people overhype its capabilities. (40:24):
undefined

Dwarkesh: I think some people are like, oh, it's already AGI, but it's like a little hobbled (40:27):
undefined

Dwarkesh: little AGI where we're like sort of giving it a concussion every couple of hours (40:31):
undefined

Dwarkesh: and like it forgets everything. (40:35):
undefined

Dwarkesh: We're like trapped in a chatbot context. But fundamentally, the thing inside (40:37):
undefined

Dwarkesh: is like a very smart human. (40:41):
undefined

Dwarkesh: I disagree with that perspective. So if that's your perspective, (40:44):
undefined

Dwarkesh: I say like, no, it's not that smart. (40:46):
undefined

Dwarkesh: Your perspective is just statistical associations. I say definitely smarter. (40:47):
undefined

Dwarkesh: Like it's like genuinely there's an intelligence there. (40:51):
undefined

Dwarkesh: And the, so one thing you could say to the person who thinks that it's already (40:54):
undefined

Dwarkesh: AGI is this, look, if a single human had as much stuff memorized as these models (40:58):
undefined

Dwarkesh: seem to have memorized, right? (41:03):
undefined

Dwarkesh: Which is to say that they have all of internet text, everything that human has (41:04):
undefined

Dwarkesh: written on the internet memorized, they would potentially be discovering all (41:08):
undefined

Dwarkesh: kinds of connections and discoveries. (41:13):
undefined

Dwarkesh: They'd notice that this thing which causes a migraine is associated with this kind of deficiency. (41:16):
undefined

Dwarkesh: So maybe if you take the supplement, your migraines will be cured. (41:22):
undefined

Dwarkesh: There'd be just this list of just like trivial connections that lead to big (41:25):
undefined

Dwarkesh: discoveries all through the place. (41:28):
undefined

Dwarkesh: It's not clear that there's been an unambiguous case of an AI just doing this by itself. (41:30):
undefined

Dwarkesh: So then why, so that's something potentially to explain, like if they're so (41:37):
undefined

Dwarkesh: intelligent, why aren't they able to use their disproportionate capabilities, (41:40):
undefined

Dwarkesh: their unique capabilities to come up with these discoveries? (41:44):
undefined

Dwarkesh: I don't think there's actually a good answer to that question yet, (41:47):
undefined

Dwarkesh: except for the fact that they genuinely aren't that creative. (41:49):
undefined

Dwarkesh: Maybe they're like intelligent in the sense of knowing a lot of things, (41:51):
undefined

Dwarkesh: but they don't have this fluid intelligence that humans have. (41:54):
undefined

Dwarkesh: Anyway, so I give you a wish-washy answer because I think some people are underselling (41:57):
undefined

Dwarkesh: the intelligence. Some people are overselling it. (42:00):
undefined

Ryan Sean Adams: I recall a tweet lately from Tyler Cowen. I think he was referring to maybe (42:03):
undefined

Ryan Sean Adams: O3, and he basically said, it feels like AGI. (42:07):
undefined

Ryan Sean Adams: I don't know if it is AGI or not, but like to me, it feels like AGI. (42:10):
undefined

Ryan Sean Adams: What do you account for this feeling of like intelligence then (42:14):
undefined

Dwarkesh: I think this is actually very interesting because it gets to a crux that Tyler (42:18):
undefined

Dwarkesh: and I have so Tyler and I disagree on two big things one he thinks you know (42:22):
undefined

Dwarkesh: as he said in the blog post 03 is AGI I don't think it's AGI I think it's, (42:28):
undefined

Dwarkesh: it's orders of magnitude less valuable or, you know, like many orders of magnitude (42:32):
undefined

Dwarkesh: less valuable and less useful than an AGI. (42:37):
undefined

Dwarkesh: That's one thing we disagree on. The other thing we disagree on is he thinks (42:39):
undefined

Dwarkesh: that once we do get AGI, we'll only see 0.5% increase in the economic growth (42:43):
undefined

Dwarkesh: rate. This is like what the internet caused, right? (42:47):
undefined

Dwarkesh: Whereas I think we will see tens of percent increase in economic growth. (42:49):
undefined

Dwarkesh: Like it will just be the difference between the pre-industrial revolution rate (42:53):
undefined

Dwarkesh: of growth versus industrial revolution, that magnitude of change again. (42:57):
undefined

Dwarkesh: And I think these two disagreements are linked because if you do believe we're (43:00):
undefined

Dwarkesh: already at AGI and you look around the world and you say like, (43:05):
undefined

Dwarkesh: well, it fundamentally looks the same, you'd be forgiven for thinking like, (43:08):
undefined

Dwarkesh: oh, there's not that much value in getting to AGI. (43:12):
undefined

Dwarkesh: Whereas if you are like me and you think like, no, we'll get this broadly at (43:14):
undefined

Dwarkesh: the minimum, at a very minimum, we'll get a broadly deployed intelligence explosion once we get to AGI, (43:17):
undefined

Dwarkesh: then you're like, OK, I'm just expecting some sort of singulitarian crazy future (43:22):
undefined

Dwarkesh: with a robot factories and, you know, solar farms all across the desert and things like that. (43:26):
undefined

Ryan Sean Adams: Yeah, I mean, it strikes me that your disagreement with Tyler is just based (43:31):
undefined

Ryan Sean Adams: on the semantic definition of like what AGI actually is. (43:35):
undefined

Ryan Sean Adams: And Tyler, it sounds like he has kind of a lower threshold for what AGI is, (43:39):
undefined

Ryan Sean Adams: whereas you have a higher threshold. (43:44):
undefined

Ryan Sean Adams: Is there like a accepted definition for AGI? (43:45):
undefined

Dwarkesh: No. One thing that's useful for the purposes of discussions is to say automating (43:48):
undefined

Dwarkesh: all white collar work because robotics hasn't made as much progress as LLMs (43:54):
undefined

Dwarkesh: have or computer use has. (43:59):
undefined

Dwarkesh: So if we just say anything a human can do or maybe 90% of what humans can do (44:01):
undefined

Dwarkesh: at a desk, an AI can also do, that's potentially a useful definition for at (44:06):
undefined

Dwarkesh: least getting the cognitive elements relevant to defining AGI. (44:11):
undefined

Dwarkesh: But yeah, there's not one definition which suits all purposes. (44:15):
undefined

Ryan Sean Adams: Do we know what's like going on inside of these models, right? (44:18):
undefined

Ryan Sean Adams: So like, you know, Josh was talking earlier in the conversation about like this (44:23):
undefined

Ryan Sean Adams: at the base being sort of token prediction, right? (44:26):
undefined

Ryan Sean Adams: And I guess this starts to raise the question of like, what is intelligence in the first place? (44:29):
undefined

Ryan Sean Adams: And these AI models, I mean, they seem like they're intelligent, (44:35):
undefined

Ryan Sean Adams: but do they have a model of the world the way maybe a human might? (44:40):
undefined

Ryan Sean Adams: Are they sort of babbling or like, is this real reasoning? (44:44):
undefined

Ryan Sean Adams: And like, what is real reasoning? Do we just judge that based on the results (44:49):
undefined

Ryan Sean Adams: or is there some way to like peek inside of its head? (44:53):
undefined

Dwarkesh: I used to have similar questions a couple of years ago. And then, (44:56):
undefined

Dwarkesh: because honestly, the things they did at the time were like ambiguous. (45:00):
undefined

Dwarkesh: You could say, oh, it's close enough to something else in this trading data set. (45:03):
undefined

Dwarkesh: That is just basically copy pasting. It didn't come up with a solution by itself. (45:07):
undefined

Dwarkesh: But we've gotten to the point where I can come up with a pretty complicated (45:12):
undefined

Dwarkesh: math problem and it will solve it. (45:17):
undefined

Dwarkesh: It can be a math problem, like not like, you know, undergrad or high school math problem. (45:19):
undefined

Dwarkesh: Like the problem we get, the problems the smartest math professors come up with (45:24):
undefined

Dwarkesh: in order to test International Math Olympiad. (45:28):
undefined

Dwarkesh: You know, the kids who spend all their life preparing for this, (45:31):
undefined

Dwarkesh: the geniuses who spend all their life, all their young adulthood preparing to (45:34):
undefined

Dwarkesh: take these really gnarly math puzzle challenges. (45:37):
undefined

Dwarkesh: And the model will get these kinds of questions, right? They require all this (45:40):
undefined

Dwarkesh: abstract creative thinking, this reasoning for hours, the model will get the right. (45:43):
undefined

Dwarkesh: Okay, so if that's not reasoning, then why is reasoning valuable again? (45:48):
undefined

Dwarkesh: Like, what exactly was this reasoning supposed to be? (45:53):
undefined

Dwarkesh: So I think they genuinely are reasoning. I mean, I think there's other capabilities (45:56):
undefined

Dwarkesh: they lack, which are actually more, in some sense, they seem to us to be more (45:59):
undefined

Dwarkesh: trivial, but actually much harder to learn. But the reasoning itself, I think, is there. (46:03):
undefined

Dwarkesh Patel: And the answer to the intelligence question is also kind of clouded, (46:07):
undefined

Dwarkesh Patel: right? Because we still really don't understand what's going on in an LLM. (46:10):
undefined

Dwarkesh Patel: Dario from Anthropoc, he recently posted the paper about interpretation. (46:14):
undefined

Dwarkesh Patel: And can you explain why we don't even really understand what's going on in these (46:17):
undefined

Dwarkesh Patel: LLMs, even though we're able to make them and yield the results from them? Mmm. (46:21):
undefined

Dwarkesh Patel: Because it very much still is kind of like a black box. We write some code, (46:27):
undefined

Dwarkesh Patel: we put some inputs in, and we get something out, but we're not sure what happens in the middle, (46:30):
undefined

Dwarkesh: Why it's creating this output. (46:34):
undefined

Dwarkesh Patel: I mean, it's exactly what you're saying. (46:36):
undefined

Dwarkesh: It's that in other systems we engineer in the world, we have to build it up bottom-ups. (46:38):
undefined

Dwarkesh: If you build a bridge, you have to understand how every single beam is contributing to the structure. (46:44):
undefined

Dwarkesh: And we have equations for why the thing will stay standing. (46:50):
undefined

Dwarkesh: There's no such thing for AI. We didn't build it, more so we grew it. (46:54):
undefined

Dwarkesh: It's like watering a plant. And a couple thousand years ago, (46:59):
undefined

Dwarkesh: they were doing agriculture, but they didn't know why. (47:03):
undefined

Dwarkesh: Why do plants grow? How do they collect energy from sunlight? All these things. (47:08):
undefined

Dwarkesh: And I think we're in a substantially similar position with respect to intelligence, (47:13):
undefined

Dwarkesh: with respect to consciousness, with respect to all these other interesting questions (47:18):
undefined

Dwarkesh: about how minds work, which is in some sense really cool because there's this (47:23):
undefined

Dwarkesh: huge intellectual horizon that's become not only available, but accessible to investigation. (47:27):
undefined

Dwarkesh: In another sense, it's scary because we know that minds can suffer. (47:33):
undefined

Dwarkesh: We know that minds have moral worth and we're creating minds and we have no (47:37):
undefined

Dwarkesh: understanding of what's happening in these minds. (47:42):
undefined

Dwarkesh: Is a process of gradient descent a painful process? (47:44):
undefined

Dwarkesh: We don't know, but we're doing a lot of it. (47:48):
undefined

Dwarkesh: So hopefully we'll learn more. But yeah, I think we're in a similar position (47:51):
undefined

Dwarkesh: to some farmer in Uruk in 3500 BC. (47:54):
undefined

Josh Kale: Wow. (47:57):
undefined

Ryan Sean Adams: And I mean, the potential, the idea that minds can suffer, minds have some moral (47:58):
undefined

Ryan Sean Adams: worth, and also minds have some free will. (48:03):
undefined

Ryan Sean Adams: They have some sort of autonomy, or maybe at least a desire to have autonomy. (48:06):
undefined

Ryan Sean Adams: I mean, this brings us to kind of this sticky subject of alignment and AI safety (48:11):
undefined

Ryan Sean Adams: and how we go about controlling the intelligence that we're creating, (48:15):
undefined

Ryan Sean Adams: if even that's what we should be doing, controlling it. And we'll get to that in a minute. (48:20):
undefined

Ryan Sean Adams: But I want to start with maybe the headlines here a little bit. (48:24):
undefined

Ryan Sean Adams: So headline just this morning, latest OpenAI models sabotaged a shutdown mechanism (48:28):
undefined

Ryan Sean Adams: despite commands to the contrary. (48:34):
undefined

Ryan Sean Adams: OpenAI's O1 model attempted to copy itself to external servers after being threatened (48:36):
undefined

Ryan Sean Adams: with shutdown that denied the action when discovered. (48:41):
undefined

Ryan Sean Adams: I've read a number of papers for this. Of course, mainstream media has these (48:44):
undefined

Ryan Sean Adams: types of headlines almost on a weekly basis now, and it's starting to get to daily. (48:48):
undefined

Ryan Sean Adams: But there does seem to be some evidence that AIs lie to us, (48:53):
undefined

Ryan Sean Adams: If that's even the right term, in order to pursue goals, goals like self-preservation, (48:58):
undefined

Ryan Sean Adams: goals like replication, even deep-seated values that we might train into them, (49:03):
undefined

Ryan Sean Adams: sort of a constitution type of value. (49:08):
undefined

Ryan Sean Adams: They seek to preserve these values, which maybe that's a good thing, (49:11):
undefined

Ryan Sean Adams: or maybe it's not a good thing if we don't actually want them to interpret the values in a certain way. (49:15):
undefined

Ryan Sean Adams: Some of these headlines that we're seeing now, To you, with your kind of corpus (49:21):
undefined

Ryan Sean Adams: of knowledge and all of the interviews and discovery you've done on your side, (49:25):
undefined

Ryan Sean Adams: is this like media sensationalism or is this like alarming? (49:29):
undefined

Ryan Sean Adams: And if it's alarming, how concerned should we be about this? (49:33):
undefined

Dwarkesh: I think on net, it's quite alarming. I do think that some of these results have (49:37):
undefined

Dwarkesh: been sort of cherry picked. (49:42):
undefined

Dwarkesh: Or if you look into the code, what's happened is basically the researchers have (49:44):
undefined

Dwarkesh: said, hey, pretend to be a bad person. (49:47):
undefined

Dwarkesh: Wow, AI is being a bad person. Isn't that crazy? (49:50):
undefined

Dwarkesh: But the system prompt is just like hey do this bad thing right now i personally (49:53):
undefined

Dwarkesh: but i have also seen other results which are not of this quality i mean the (49:57):
undefined

Dwarkesh: the clearest example so backing up, (50:02):
undefined

Dwarkesh: what is the reason to think this will be a bigger problem in the future than (50:05):
undefined

Dwarkesh: it is now because we all interact with these systems and they're actually like (50:08):
undefined

Dwarkesh: quite moral or aligned right like you can talk to a chatbot and you like ask (50:13):
undefined

Dwarkesh: it to how should you deal with some crisis where there's a correct answer, (50:17):
undefined

Dwarkesh: you know, like it will tell you not to be violent. It'll give you reasonable advice. (50:22):
undefined

Dwarkesh: It seems to have good values. So it's worth noticing this, right? (50:26):
undefined

Dwarkesh: And being happy about it. (50:29):
undefined

Dwarkesh: The concern is that we're moving from a regime where we've trained them on human (50:31):
undefined

Dwarkesh: language, which implicitly has human morals and the way, you know, (50:36):
undefined

Dwarkesh: normal people think about values implicit in it. (50:41):
undefined

Dwarkesh: Plus this RLHF process we did to a regime where we're mostly spending compute (50:44):
undefined

Dwarkesh: on just having them answer problems yes or no or correct or not rather just like. (50:50):
undefined

Dwarkesh: And pass all the unit tests, get the right answer on this math problem. (50:58):
undefined

Dwarkesh: And this has no guardrails intrinsically in terms of what is allowed to do, (51:02):
undefined

Dwarkesh: what is the proper moral way to do something. (51:09):
undefined

Dwarkesh: I think that can be a loaded term, but here's a more concrete example. (51:11):
undefined

Dwarkesh: One problem we're running into with these coding agents more and more, (51:15):
undefined

Dwarkesh: and this has nothing to do with these abstract concerns about alignment, (51:18):
undefined

Dwarkesh: but more so just like how do we get economic value out of these models, (51:21):
undefined

Dwarkesh: is that Claude or Gemini will, instead of writing code such that it passes the unit tests, (51:24):
undefined

Dwarkesh: it will often just delete the unit tests so that the code just passes by default. (51:32):
undefined

Dwarkesh: Now, why would it do that? Well, it's learned in the process. (51:37):
undefined

Dwarkesh: It was trained on the goal during training of you must pass all unit tests. (51:41):
undefined

Dwarkesh: And probably within some environment in which it was trained, (51:45):
undefined

Dwarkesh: it was able to just get away. (51:48):
undefined

Dwarkesh: Like there wasn't designed well enough. And so it found this like little hole (51:49):
undefined

Dwarkesh: where it could just like delete the file that had the unit test or rewrite them (51:53):
undefined

Dwarkesh: so that it always said, you know, equals true, then pass. (51:55):
undefined

Dwarkesh: And right now we can discover these even without, even though we can discover (51:59):
undefined

Dwarkesh: these, you know, it's still past, there's still been enough hacks like this, (52:03):
undefined

Dwarkesh: such that the model is like becoming more and more hacky like that. (52:06):
undefined

Dwarkesh: In the future, we're going to be training models in ways that we is beyond our (52:10):
undefined

Dwarkesh: ability to even understand, certainly beyond everybody's ability to understand. (52:14):
undefined

Dwarkesh: There may be a few people who might be able to see just the way that right now, (52:18):
undefined

Dwarkesh: if you came up with a new math proof for some open problem in mathematics, (52:21):
undefined

Dwarkesh: there will be only be a few people in the world who will be able to evaluate that math proof. (52:24):
undefined

Dwarkesh: We'll be in a similar position with respect to all of the things that these (52:28):
undefined

Dwarkesh: models are being trained on at the frontier, especially math and code, (52:31):
undefined

Dwarkesh: because humans were big dum-dums with respect to this reasoning stuff. (52:34):
undefined

Dwarkesh: And so there's a sort of like first principles reason to expect that this new (52:38):
undefined

Dwarkesh: modality of training will be less amenable to the kinds of supervision that (52:41):
undefined

Dwarkesh: was grounded within the pre-training corpus. (52:46):
undefined

Ryan Sean Adams: I don't know that everyone has kind of an intuition or an idea why it doesn't (52:49):
undefined

Ryan Sean Adams: work to just say like, so if we don't want our AI models to lie to us, (52:54):
undefined

Ryan Sean Adams: why can't we just tell them not to lie? (52:59):
undefined

Ryan Sean Adams: Why can't we just put that as part of their core constitution? (53:01):
undefined

Ryan Sean Adams: If we don't want our AI models to be sycophants, why can't we just say, (53:05):
undefined

Ryan Sean Adams: hey, if I tell you I want the truth, not to flatter me, just give me the straight up truth. (53:10):
undefined

Ryan Sean Adams: Why is this even difficult to do? (53:16):
undefined

Dwarkesh: Well, fundamentally, it comes down to how we train them. And we don't know how (53:18):
undefined

Dwarkesh: to train them in a way that does not reward lying or sycophancy. (53:22):
undefined

Dwarkesh: In fact, the problem is OpenAI, they explained why their recent model of theirs (53:26):
undefined

Dwarkesh: was they had to take down was just sycophantic. (53:30):
undefined

Dwarkesh: And the reason was just that they rolled out, did it in the A-B test and the (53:33):
undefined

Dwarkesh: version, the test that was more sycophantic was just preferred by users more. (53:37):
undefined

Dwarkesh: Sometimes you prefer the lie. (53:42):
undefined

Dwarkesh: Yeah, so that's, if that's what's preferred in training, you know, (53:44):
undefined

Dwarkesh: Or, for example, in the context of lying, if we've just built RL environments (53:47):
undefined

Dwarkesh: in which we're training these models, where they're going to be more successful if they lie, right? (53:52):
undefined

Dwarkesh: So if they delete the unit tests and then tell you, I passed this program and (53:59):
undefined

Dwarkesh: all the unit tests have succeeded, it's like lying to you, basically. (54:06):
undefined

Dwarkesh: And if that's what is rewarded in the process of gradient descent, (54:09):
undefined

Dwarkesh: then it's not surprising that the model you interact with will just have this (54:12):
undefined

Dwarkesh: drive to lie if it gets it closer to its goal. (54:17):
undefined

Dwarkesh: And I would just expect this to keep happening unless we can solve this fundamental (54:20):
undefined

Dwarkesh: problem that comes about in training. (54:24):
undefined

Dwarkesh Patel: So you mentioned how like ChatGPT had a version that was sycophantic, (54:26):
undefined

Dwarkesh Patel: and that's because users actually wanted that. (54:30):
undefined

Dwarkesh Patel: Who is in control? Who decides the actual alignment of these models? (54:32):
undefined

Dwarkesh Patel: Because users are saying one thing, and then they deploy it, (54:36):
undefined

Dwarkesh Patel: and then it turns out that's not actually what people want. (54:38):
undefined

Dwarkesh Patel: How do you kind of form consensus around this alignment or these alignment principles? (54:41):
undefined

Dwarkesh: Right now, obviously, it's the labs who decided this, right? (54:47):
undefined

Dwarkesh: And the safety teams of the labs. (54:49):
undefined

Dwarkesh: And I guess the question you could ask is then who should decide these? Because this will be... (54:51):
undefined

Dwarkesh Patel: Assuming the trajectory, yeah. So we keep going to get more powerful. (54:56):
undefined

Dwarkesh: Because this will be the key modality that all of us use to get, (54:59):
undefined

Dwarkesh: not only get work done, but even like, I think at some point, (55:03):
undefined

Dwarkesh: a lot of people's best friends will be AIs, at least functionally in the sense (55:06):
undefined

Dwarkesh: of who do they spend the most amount of time talking to. It might already be AIs. (55:10):
undefined

Dwarkesh: This will be the key layer in your business that you're using to get work done (55:14):
undefined

Dwarkesh: so this process of training which shapes their personality who gets to control (55:20):
undefined

Dwarkesh: it I mean it will be the laughs functionally, (55:25):
undefined

Dwarkesh: But maybe you mean, like, who should control it, right? I honestly don't know. (55:30):
undefined

Dwarkesh: I mean, I don't know if there's a better alternative to the labs. (55:34):
undefined

Dwarkesh Patel: Yeah, I would assume, like, there's some sort of social consensus, (55:36):
undefined

Dwarkesh Patel: right? Similar to how we have in America, the Constitution. (55:39):
undefined

Dwarkesh Patel: There's, like, this general form of consensus that gets formed around how we (55:41):
undefined

Dwarkesh Patel: should treat these models as they become as powerful as we think they probably will be. (55:44):
undefined

Dwarkesh: Honestly, I don't have, I don't know if anybody has a good answer about how (55:47):
undefined

Dwarkesh: you do this process. I think we lucked out, we just, like, really lucked out with the Constitution. (55:49):
undefined

Dwarkesh: It also wasn't a democratic process which resulted in the constitution, (55:55):
undefined

Dwarkesh: even though it instituted a Republican form of government. (55:58):
undefined

Dwarkesh: It was just delegates from each state. They haggled it out over the course of a few months. (56:00):
undefined

Dwarkesh: Maybe that's what happens with AI. But is there some process which feels both (56:05):
undefined

Dwarkesh: fair and which will result in actually a good constitution for these AIs? (56:10):
undefined

Dwarkesh: It's not obvious to me that, I mean, nothing comes up to the top of my head. (56:15):
undefined

Dwarkesh: Like, oh, this, you know, do rank choice voting or something. (56:19):
undefined

Dwarkesh Patel: Yeah, so I was going to ask, is there any, I mean, having spoken to everyone (56:22):
undefined

Dwarkesh Patel: who you've spoken to is there any alignment path which looks most promising which (56:24):
undefined

Dwarkesh: Feels the. (56:27):
undefined

Dwarkesh Patel: Most comforting and exciting to you (56:28):
undefined

Dwarkesh: I i think alignment in the sense of you (56:30):
undefined

Dwarkesh: know and eventually we'll have these super intelligent systems what do we do (56:33):
undefined

Dwarkesh: about that i think the the approach that i think is most promising is less about (56:36):
undefined

Dwarkesh: finding some holy grail some you know giga brain solution some equation which (56:44):
undefined

Dwarkesh: solves the whole puzzle and more like one. (56:49):
undefined

Dwarkesh: Having this Swiss cheese approach where, look, we kind of have gotten really good at jailbreaks. (56:53):
undefined

Dwarkesh: I'm sure you've heard a lot about jailbreaks over the last few years. (57:01):
undefined

Dwarkesh: It's actually much harder to jailbreak these models because, (57:03):
undefined

Dwarkesh: you know, people try to whack at these things in different ways. (57:06):
undefined

Dwarkesh: Model developers just like patched these obvious ways to do jailbreaks. (57:10):
undefined

Dwarkesh: The model also got smarter. So it's better able to understand when somebody (57:14):
undefined

Dwarkesh: is trying to jailbreak into it. (57:18):
undefined

Dwarkesh: That, I think, is one approach. Another is, I think, competition. (57:20):
undefined

Dwarkesh: I think the scary version of the future is where you have this dynamic where (57:24):
undefined

Dwarkesh: a single model and its copies are controlling the entire economy. (57:27):
undefined

Dwarkesh: When politicians want to understand what policies to pass, they're only talking (57:31):
undefined

Dwarkesh: to copies of a single model. (57:35):
undefined

Dwarkesh: If there's multiple different AI companies who are at the frontier, (57:36):
undefined

Dwarkesh: who have competing services, and whose models can monitor each other, right? (57:40):
undefined

Dwarkesh: So Claude may care about its own copies being successful in the world and it (57:44):
undefined

Dwarkesh: might be able to willing to lie on their behalf, even if you ask one copy to supervise another. (57:50):
undefined

Dwarkesh: I think you get some advantage from a copy of OpenAI's model monitoring a copy (57:53):
undefined

Dwarkesh: of DeepSeek's model, which actually brings us back to the Constitution, right? (57:58):
undefined

Dwarkesh: One of the most brilliant things in the Constitution is the system of checks and balances. (58:01):
undefined

Dwarkesh: So some combination of the Swiss cheese approach to model development and training (58:04):
undefined

Dwarkesh: and alignment, where you're careful, if you notice this kind of reward hacking, (58:09):
undefined

Dwarkesh: you do your best to solve it. (58:13):
undefined

Dwarkesh: You try to keep as much of the models thinking in human language rather than (58:14):
undefined

Dwarkesh: letting it think in AI thought in this latent space thinking. (58:19):
undefined

Dwarkesh: And the other part of it is just having normal market competition between these (58:23):
undefined

Dwarkesh: companies so that you can use them to check each other and no one company or (58:27):
undefined

Dwarkesh: no one AI is dominating the economy or advisory roles for governments. (58:31):
undefined

Ryan Sean Adams: I really like this like bundle of ideas that you sort of put together in that (58:41):
undefined

Ryan Sean Adams: because like, I think a lot of the, you know, AI safety conversation is always (58:45):
undefined

Ryan Sean Adams: couched in terms of control. (58:50):
undefined

Ryan Sean Adams: Like we have to control the thing that is the way. And I always get a little (58:52):
undefined

Ryan Sean Adams: worried when I hear like terms like control. (58:56):
undefined

Ryan Sean Adams: And it reminds me of a blog post I think you put out, which I'm hopeful you continue to write on. (58:59):
undefined

Ryan Sean Adams: I think you said it was going to be like one of a series, which is this idea (59:05):
undefined

Ryan Sean Adams: of like classical liberal AGI. And we were talking about themes like balance of power. (59:08):
undefined

Ryan Sean Adams: Let's have Claude check in with ChatGPT and monitor it. (59:13):
undefined

Josh Kale: When you have themes like transparency as well, (59:17):
undefined

Ryan Sean Adams: That feels a bit more, you know, classically liberal coded than maybe some of (59:19):
undefined

Ryan Sean Adams: the other approaches that I've heard. (59:25):
undefined

Ryan Sean Adams: And you wrote this in the post, which I thought was kind of, (59:27):
undefined

Ryan Sean Adams: it just sparked my interest because I'm not sure where you're going to go next (59:30):
undefined

Ryan Sean Adams: with this, but you said the most likely way this happens, (59:33):
undefined

Ryan Sean Adams: that is AIs have a stake in humanity's future, is if it's in the AI's best interest (59:37):
undefined

Ryan Sean Adams: to operate within our existing laws and norms. (59:42):
undefined

Ryan Sean Adams: You know, this whole idea that like, hey, the way to get true AI alignment is (59:45):
undefined

Ryan Sean Adams: to make it easy, make it the path of least resistance for AI to basically partner with humans. (59:49):
undefined

Ryan Sean Adams: It's almost this idea if the aliens (59:56):
undefined

Ryan Sean Adams: landed or something, we would create treaties with the aliens, right? (59:59):
undefined

Ryan Sean Adams: We would want them to adopt our norms. We would want to initiate trade with them. (01:00:03):
undefined

Ryan Sean Adams: Our first response shouldn't be, let's try to dominate and control them. (01:00:08):
undefined

Ryan Sean Adams: Maybe it should be, let's try to work with them. Let's try to collaborate. (01:00:13):
undefined

Ryan Sean Adams: Let's try to open up trade. (01:00:16):
undefined

Ryan Sean Adams: What's your idea here? And like, are you planning to write further posts about this? (01:00:18):
undefined

Dwarkesh: Yeah, I want to. It's just such a hard topic to think about that, (01:00:22):
undefined

Dwarkesh: you know, something always comes up. (01:00:25):
undefined

Dwarkesh: But the fundamental point I was making is, look, in the long run, (01:00:26):
undefined

Dwarkesh: if AIs are, you know, human labor is going to be obsolete because of these inherent (01:00:32):
undefined

Dwarkesh: advantages that digital minds will have and robotics will eventually be solved. (01:00:37):
undefined

Dwarkesh: So our only leverage on the future will no longer come from our labor. (01:00:41):
undefined

Dwarkesh: It will come from our legal and economic control over the society that AIs will (01:00:50):
undefined

Dwarkesh: be participating in, right? So, you know, AIs might make the economy explode (01:00:58):
undefined

Dwarkesh: in the sense of grow a lot. (01:01:03):
undefined

Dwarkesh: And for humans to benefit from that, it would have to be the case that AIs still (01:01:04):
undefined

Dwarkesh: respect your equity in the S&P 500 companies that you bought, right? (01:01:08):
undefined

Dwarkesh: Or for the AIs to follow your laws, which say that you can't do violence onto (01:01:13):
undefined

Dwarkesh: humans and you got to respect humans' properties. (01:01:18):
undefined

Josh Kale: It would have to be the case that AIs are actually bought into our (01:01:21):
undefined

Dwarkesh: System of government, into our laws and norms. And for that to happen, (01:01:25):
undefined

Dwarkesh: the way that likely happens is if it's just like the default path for the AIs (01:01:30):
undefined

Dwarkesh: as they're getting smarter and they're developing their own systems of enforcement (01:01:37):
undefined

Dwarkesh: and laws to just participate in human laws and governments. (01:01:41):
undefined

Dwarkesh: And the metaphor I use here is right now you pay half your paycheck in taxes, (01:01:46):
undefined

Dwarkesh: probably half of your taxes in some way just go to senior citizens, right? (01:01:53):
undefined

Dwarkesh: Medicare and Social Security and other programs like this. (01:01:59):
undefined

Dwarkesh: And it's not because you're in some deep moral sense aligned with senior citizens. (01:02:04):
undefined

Dwarkesh: It's not like you're spending all your time thinking about like, (01:02:09):
undefined

Dwarkesh: my main priority in life is to earn money for senior citizens. (01:02:11):
undefined

Dwarkesh: It's just that you're not going to overthrow the government to get out of paying this tax. And so... (01:02:15):
undefined

Ryan Sean Adams: Also, I happen to like my grandmother. She's fantastic. You know, (01:02:22):
undefined

Ryan Sean Adams: it's those reasons too. But yeah. (01:02:25):
undefined

Dwarkesh: So that's why you give money to your grandmother directly. But like, (01:02:26):
undefined

Dwarkesh: why are you giving money to some retiree in Illinois? Yes. (01:02:29):
undefined

Josh Kale: Yes. (01:02:33):
undefined

Dwarkesh: Yeah, it's like, okay, you could say it's like, sometimes people, (01:02:34):
undefined

Dwarkesh: some people are trying to that post by saying like, oh no, I like deeply care (01:02:37):
undefined

Dwarkesh: about the system of social welfare. (01:02:39):
undefined

Dwarkesh: I'm just like, okay, maybe you do, but I don't think like the average person (01:02:41):
undefined

Dwarkesh: is giving away hundreds of thousands of dollars a year, tens of thousands of (01:02:45):
undefined

Dwarkesh: dollars a year to like some random stranger they don't know, (01:02:47):
undefined

Dwarkesh: who's like, who's not like especially in need of charity, right? (01:02:50):
undefined

Dwarkesh: Like most senior citizens have some savings. (01:02:53):
undefined

Dwarkesh: It's just, it's just because this is a law and you like, you give it to them (01:02:55):
undefined

Dwarkesh: or you'll get, go to jail. (01:02:59):
undefined

Dwarkesh: But fundamentally, if the tax was like 99%, you would, like, (01:03:01):
undefined

Dwarkesh: you would, maybe you wouldn't overthrow the government. You'd just, (01:03:05):
undefined

Dwarkesh: like, leave the jurisdiction. (01:03:07):
undefined

Dwarkesh: You'd, like, emigrate somewhere. And AIs can potentially also do this, (01:03:08):
undefined

Dwarkesh: right? There's more than one country. (01:03:12):
undefined

Dwarkesh: They could, like, there's countries which would be more AI forward. (01:03:14):
undefined

Dwarkesh: And it would be a bad situation to end up in where... (01:03:16):
undefined

Dwarkesh: All this explosion in AI technology is happening in the country, (01:03:21):
undefined

Dwarkesh: which is doing the least amount to protect humans', (01:03:24):
undefined

Dwarkesh: rights and to provide some sort of monetary compensation to humans once their (01:03:28):
undefined

Dwarkesh: labor is no longer valuable. (01:03:36):
undefined

Dwarkesh: So our labor could be worth nothing, but because of how much richer the world (01:03:38):
undefined

Dwarkesh: is after AI, you have these billions of extra researchers, workers, etc. (01:03:42):
undefined

Dwarkesh: It could still be trivial to have individual humans have the equivalent of millions, (01:03:47):
undefined

Dwarkesh: even billions of dollars worth of wealth. In fact, it might literally be invaluable (01:03:54):
undefined

Dwarkesh: amounts of wealth in the following sense. So here's an interesting thought experiment. (01:03:59):
undefined

Dwarkesh: Imagine you have this choice. You can go back to the year 1500, (01:04:02):
undefined

Dwarkesh: but you know, of course, the year 1500 kind of sucks. (01:04:07):
undefined

Dwarkesh: You have no antibiotics, no TV, no running water. But here's how I'll make it up to you. (01:04:09):
undefined

Dwarkesh: I can give you any amount of money, but you can only use that amount of money in the year 1500. (01:04:15):
undefined

Dwarkesh: And you'll go back with these sacks of gold. How much money would I have to (01:04:20):
undefined

Dwarkesh: give you that you can use in the year 1500 to make you go back? And plausibly. (01:04:24):
undefined

Dwarkesh Patel: The answer is (01:04:27):
undefined

Dwarkesh: There's no amount of money you would rather have in the year 1500 than just (01:04:27):
undefined

Dwarkesh: have a normal life today. (01:04:30):
undefined

Dwarkesh: And we could be in a similar position with regards to the future where there's (01:04:31):
undefined

Dwarkesh: all these different, I mean, you'll have much better health, (01:04:36):
undefined

Dwarkesh: like physical health, mental health, longevity. (01:04:39):
undefined

Dwarkesh: That's just like the thing we can contemplate now. But people in 1500 couldn't (01:04:42):
undefined

Dwarkesh: contemplate the kinds of quality of life advances we would have 500 years later, (01:04:45):
undefined

Dwarkesh: right? So anyways, this is all to say that this could be our future for humans, (01:04:49):
undefined

Dwarkesh: even if our labor isn't worth anything. (01:04:54):
undefined

Dwarkesh: But it does require us to have AIs that choose to participate or in some way (01:04:57):
undefined

Dwarkesh: incentivize to participate in some system which we have leverage over. (01:05:05):
undefined

Ryan Sean Adams: Yeah, I find this just such a fast, I'm hopeful we do some more exploration (01:05:12):
undefined

Ryan Sean Adams: around this because I think what you're calling for is basically like, (01:05:16):
undefined

Ryan Sean Adams: what you would be saying is invite them into our property rights system. (01:05:19):
undefined

Ryan Sean Adams: I mean, there are some that are calling in order to control AI, (01:05:22):
undefined

Ryan Sean Adams: they have great power, but they don't necessarily have capabilities. (01:05:25):
undefined

Ryan Sean Adams: So we shouldn't allow AI to hold money or to have property. (01:05:28):
undefined

Ryan Sean Adams: I think you would say, no, actually, the path forward to alignment is allow (01:05:31):
undefined

Ryan Sean Adams: AI to have some vested interest in our property rights system and some stake (01:05:36):
undefined

Ryan Sean Adams: in our governance, potentially, right? (01:05:42):
undefined

Ryan Sean Adams: The ability to vote, almost like a constitution for AIs. (01:05:44):
undefined

Ryan Sean Adams: I'm not sure how this would work, but it's a fascinating thought experiment. (01:05:47):
undefined

Dwarkesh: I will say one thing I think this could end disastrously if we give them a stake (01:05:53):
undefined

Dwarkesh: in their property system but we let them play, (01:06:00):
undefined

Dwarkesh: us off each other. So if you think about, there's many cases in history where (01:06:04):
undefined

Dwarkesh: the British, initially, the East India Trading Company was genuinely a trading (01:06:09):
undefined

Dwarkesh: company that operated in India. (01:06:13):
undefined

Dwarkesh: And it was able to play off, you know, it was like doing trade with different, (01:06:15):
undefined

Dwarkesh: different, you know, provinces in India, there was no single powerful leader. (01:06:18):
undefined

Dwarkesh: And by playing, you know, by doing trade, one of them, leveraging one of their (01:06:23):
undefined

Dwarkesh: armies, etc., they were able to conquer the continent. Similar thing could happen to human society. (01:06:28):
undefined

Dwarkesh: The way to avoid such an outcome at a high level is involves us playing the (01:06:32):
undefined

Dwarkesh: AIs off each other instead, right? (01:06:38):
undefined

Dwarkesh: So this is why I think competition is such a big part of the puzzle, (01:06:40):
undefined

Dwarkesh: having different AIs monitor each other, having this bargaining position where (01:06:45):
undefined

Dwarkesh: there's not just one company that's at the frontier. (01:06:49):
undefined

Dwarkesh: Another example here is if you think about how the Spanish conquered all these (01:06:51):
undefined

Dwarkesh: new world empires, it's actually so crazy that a couple hundred conquistaDwars (01:06:55):
undefined

Dwarkesh: would show up and conquer a nation of 10 million people, the Incas, (01:06:58):
undefined

Dwarkesh: Aztecs. And why were they able to do this? (01:07:03):
undefined

Dwarkesh: Well, one of the reasons is the Spanish were able to learn from each of their (01:07:05):
undefined

Dwarkesh: previous expeditions, whereas the Native Americans were not. (01:07:11):
undefined

Dwarkesh: So Cortez learned from how Cuba was subjugated when he conquered the Aztecs. (01:07:15):
undefined

Dwarkesh: Pizarro was able to learn from how Cortez conquered the Aztecs when he conquered the Incas. (01:07:22):
undefined

Dwarkesh: The Incas didn't even know the Aztecs existed. So eventually there was this (01:07:25):
undefined

Dwarkesh: uprising against Pizarro and Manco Inca led an insurgency where they actually (01:07:30):
undefined

Dwarkesh: did figure out how to fight horses, (01:07:36):
undefined

Dwarkesh: how to fight people, you know, people in armor on horses, don't fight them on (01:07:37):
undefined

Dwarkesh: flat terrain, throw rocks down at them, et cetera. (01:07:42):
undefined

Dwarkesh: But by this point, it was too late. If they knew this going into the battle, (01:07:44):
undefined

Dwarkesh: the initial battle, they might've been able to fend off because, (01:07:48):
undefined

Dwarkesh: you know, just as the conquistaDwars only arrived at a few hundred soldiers, (01:07:51):
undefined

Dwarkesh: we're going to the age of AI with a tremendous amount of leverage. (01:07:54):
undefined

Dwarkesh: We literally control all the stuff, right? (01:07:58):
undefined

Dwarkesh: But we just need to lock in our advantage. We just need to be in a position (01:08:01):
undefined

Dwarkesh: where, you know, they're not going to be able to play us off each other. (01:08:04):
undefined

Dwarkesh: We're going to be able to learn what their weaknesses are. (01:08:08):
undefined

Dwarkesh: And this is why I think one good idea, for example, would be that, (01:08:11):
undefined

Dwarkesh: look, DeepSeek is a Chinese company. (01:08:14):
undefined

Dwarkesh: It would be good if, suppose DeepSeek did something naughty, (01:08:17):
undefined

Dwarkesh: like the kinds of experiments we're talking about right now where it hacks the (01:08:21):
undefined

Dwarkesh: unit tests or so forth. I mean, eventually these things will really matter. (01:08:24):
undefined

Dwarkesh: Like Xi Jinping is listening to AIs because they're so smart and they're so capable. (01:08:27):
undefined

Dwarkesh: If China notices that their AIs are doing something bad, or they notice a failed (01:08:32):
undefined

Dwarkesh: coup attempt, for example, (01:08:37):
undefined

Dwarkesh: it's very important that they tell us And we tell them if we notice something (01:08:38):
undefined

Dwarkesh: like that on our end, it would be like the Aztecs and Incas talking to each (01:08:43):
undefined

Dwarkesh: other about like, you know, this is what happens. (01:08:46):
undefined

Dwarkesh: This is how you fight. This is how you fight horses. (01:08:49):
undefined

Dwarkesh: This is the kind of tactics and deals they try to make with you. Don't trust them, etc. (01:08:51):
undefined

Dwarkesh: It would require cooperation on humans' part to have this sort of red telephone. (01:08:56):
undefined

Dwarkesh: So during the Cold War, there was this red telephone between America and the (01:08:59):
undefined

Dwarkesh: Soviet Union after the human missile crisis, where just to make sure there's (01:09:03):
undefined

Dwarkesh: no misunderstandings, they're like, okay, if we think something's going on, (01:09:06):
undefined

Dwarkesh: let's just hop on the call. (01:09:08):
undefined

Dwarkesh: I think we should have a similar policy with respect to these kinds of initial (01:09:10):
undefined

Dwarkesh: warning signs we'll get from AI so that we can learn from each other. (01:09:15):
undefined

Dwarkesh Patel: Awesome. Okay, so now that we've described this artificial gender intelligence, (01:09:19):
undefined

Dwarkesh Patel: I want to talk about how we actually get there. How do we build it? (01:09:22):
undefined

Dwarkesh Patel: And a lot of this we've been discussing kind of takes place in this world of (01:09:25):
undefined

Dwarkesh Patel: bits. But you have this great chapter in the book called Inputs, (01:09:27):
undefined

Dwarkesh Patel: which discusses the physical world around us, where you can't just write a few strings of code. (01:09:30):
undefined

Dwarkesh Patel: You actually have to go and move some dirt and you have to ship servers places (01:09:35):
undefined

Dwarkesh Patel: and you need to power it and you need physical energy from meat space. (01:09:38):
undefined

Dwarkesh Patel: And you kind of describe these limiting factors where we have compute, (01:09:41):
undefined

Dwarkesh Patel: we have energy, we have data. (01:09:46):
undefined

Dwarkesh Patel: What I'm curious to know is, do we have enough of this now? or is there a clear (01:09:47):
undefined

Dwarkesh Patel: path to get there in order to build the AGI? (01:09:52):
undefined

Dwarkesh Patel: Basically, what needs to happen in order for us to get to this place that you're describing? (01:09:54):
undefined

Dwarkesh: We only have a couple more years left of this scaling, (01:09:57):
undefined

Dwarkesh: this exponential scaling before we're hitting these inherent roadblocks of energy (01:10:02):
undefined

Dwarkesh: and our ability to manufacture ships, which means that if scaling is going to (01:10:09):
undefined

Dwarkesh: work to deliver us AGI, it has to work by 2028. (01:10:13):
undefined

Dwarkesh: Otherwise, we're just left with mostly algorithmic progress, (01:10:17):
undefined

Dwarkesh: But even within algorithmic progress, the sort of low-hanging fruit in this (01:10:19):
undefined

Dwarkesh: deep learning paradigm is getting more and more plucked. (01:10:23):
undefined

Dwarkesh: So then the odds per year of getting to AGI diminish a lot, right? (01:10:26):
undefined

Dwarkesh: So there is this weird, funny thing happening right now where we either discover (01:10:31):
undefined

Dwarkesh: AGI within the next few years, (01:10:36):
undefined

Dwarkesh: or the yearly probability craters, and then we might be looking at decades of (01:10:40):
undefined

Dwarkesh: further research that's required in terms of algorithms to get to AGI. (01:10:44):
undefined

Dwarkesh: I am of the opinion that some algorithmic progress is necessarily needed because (01:10:47):
undefined

Dwarkesh: there's no easy way to solve continual learning just by making the context length (01:10:51):
undefined

Dwarkesh: bigger or just by doing RL. (01:10:55):
undefined

Dwarkesh: That being said, I just think the progress so far has been so remarkable that, (01:10:57):
undefined

Dwarkesh: you know, 2032 is very close. (01:11:01):
undefined

Dwarkesh: My time has to be slightly longer than that, but I think it's extremely plausible (01:11:04):
undefined

Dwarkesh: that we're going to see a broadly deployed intelligence explosion within the next 10 years. (01:11:08):
undefined

Dwarkesh Patel: And one of these key inputs is energy, right? a lot, I actually heard it mentioned (01:11:13):
undefined

Dwarkesh Patel: on your podcast, is the United States relative to China on this particular place (01:11:18):
undefined

Dwarkesh Patel: of energy, where China is adding, what is the stat? (01:11:24):
undefined

Dwarkesh Patel: I think it's one United States worth of energy every 18 months. (01:11:26):
undefined

Dwarkesh Patel: And their plan is to go from three to eight terawatts of power versus the United (01:11:30):
undefined

Dwarkesh Patel: States, one to two terawatts of power by 2030. (01:11:34):
undefined

Dwarkesh Patel: So given that context of that one resource alone, is China better equipped to (01:11:36):
undefined

Dwarkesh Patel: get to that place versus with the United States? (01:11:41):
undefined

Dwarkesh: So right now, America has a big advantage in terms of chips. (01:11:44):
undefined

Dwarkesh: China doesn't have the ability to manufacture leading-edge semiconductors, (01:11:50):
undefined

Dwarkesh: and these are the chips that go into... (01:11:53):
undefined

Dwarkesh: You need these dyes in order to have the kinds of AI chips to... (01:11:56):
undefined

Dwarkesh: You need millions of them in order to have a frontier AI system. (01:12:01):
undefined

Dwarkesh: Eventually, China will catch up in this arena as well, right? (01:12:08):
undefined

Dwarkesh: Their technology will catch up. So the export controls will keep us ahead in (01:12:10):
undefined

Dwarkesh: this category for 5, 10 years. (01:12:14):
undefined

Dwarkesh: But if we're looking in the world where timelines are long, which is to say (01:12:16):
undefined

Dwarkesh: that AGI isn't just right around the corner, they will have this overwhelming (01:12:19):
undefined

Dwarkesh: energy advantage and they'll have caught up in chips. (01:12:23):
undefined

Dwarkesh: So then the question is like, why wouldn't they win at that point? (01:12:27):
undefined

Dwarkesh: So the longer you think we're away from AGI, the more it looks like China's game to lose. (01:12:30):
undefined

Dwarkesh: I mean, if you look in the nitty gritty, I think it's more about having centralized (01:12:37):
undefined

Dwarkesh: sources of power because you need to train the AI in one place. (01:12:42):
undefined

Dwarkesh: This might be changing with RL, but it's very important to have a single site (01:12:47):
undefined

Dwarkesh: which has a gigawatt, two gigawatts more power. (01:12:51):
undefined

Dwarkesh: And if we ramped up natural gas, you know, you can get generators and natural (01:12:54):
undefined

Dwarkesh: gas and maybe it's possible to do a last ditch effort, even if our overall energy (01:13:00):
undefined

Dwarkesh: as a country is lower than China's. The question is whether we will have the (01:13:04):
undefined

Dwarkesh: political will to do that. (01:13:07):
undefined

Dwarkesh: I think people are sort of underestimating how much of a backlash there will be against AI. (01:13:08):
undefined

Dwarkesh: The government needs to make proactive efforts in order to make sure that America (01:13:14):
undefined

Dwarkesh: stays at the leading edge in AI from zoning of data centers to how copyright (01:13:18):
undefined

Dwarkesh: is handled for data for these models. (01:13:25):
undefined

Dwarkesh: And if we mess up, if it becomes too hard to develop in America, (01:13:27):
undefined

Dwarkesh: I think it would genuinely be China's game to lose. (01:13:32):
undefined

Ryan Sean Adams: And do you think this narrative is right, that whoever wins the AGI war, (01:13:34):
undefined

Ryan Sean Adams: kind of like whoever gets to AGI first, just basically wins the 21st century? Is it that simple? (01:13:38):
undefined

Dwarkesh: I don't think it's just a matter of training the frontier system. (01:13:43):
undefined

Dwarkesh: I think people underestimate how important it is to have the compute available to run these systems. (01:13:46):
undefined

Dwarkesh: Because eventually once you get to AGI, just think of it like a person. (01:13:51):
undefined

Dwarkesh: And what matters then is how many people you have. (01:13:55):
undefined

Dwarkesh: I mean, it actually is the main thing that matters today as well, (01:13:59):
undefined

Dwarkesh: right? Like, why could China take over Taiwan if it wanted to? (01:14:02):
undefined

Dwarkesh: And if it didn't have America, you know, America, it didn't think America would intervene. (01:14:05):
undefined

Dwarkesh: But because Taiwan has 20 million people or on the order of 20 million people (01:14:09):
undefined

Dwarkesh: and China has 1.4 billion people. (01:14:13):
undefined

Dwarkesh: You could have a future where if China has way more compute than us, (01:14:17):
undefined

Dwarkesh: but equivalent levels of AI, it would be like the relationship between China and Taiwan. (01:14:21):
undefined

Dwarkesh: Their population is functionally so much higher. This just means more research, (01:14:27):
undefined

Dwarkesh: more factories, more development, more ideas. (01:14:31):
undefined

Dwarkesh: So this inference capacity, this capacity to deploy AIs will actually probably (01:14:35):
undefined

Dwarkesh: be the thing that determines who wins the 21st century. (01:14:41):
undefined

Ryan Sean Adams: So this is like the scaling law applied to, I guess, nation state geopolitics, right? (01:14:44):
undefined

Ryan Sean Adams: And it's back to compute plus data wins. (01:14:50):
undefined

Ryan Sean Adams: If compute plus data wins superintelligence, compute plus data also wins geopolitics. (01:14:54):
undefined

Dwarkesh: Yep. And the thing to be worried about is that China, speaking of compute plus (01:15:00):
undefined

Dwarkesh: data, China also has a lot more data on the real world, right? (01:15:04):
undefined

Dwarkesh: If you've got entire megalopolises filled with factories where you're already (01:15:08):
undefined

Dwarkesh: deploying robots and different production systems which use automation, (01:15:13):
undefined

Dwarkesh: you have in-house this process knowledge you're building up which the AIs can (01:15:19):
undefined

Dwarkesh: then feed on and accelerate. (01:15:24):
undefined

Dwarkesh: That equivalent level of data we don't have in America. (01:15:27):
undefined

Dwarkesh: So this could be a period in which those technological advantages or those advantages (01:15:31):
undefined

Dwarkesh: in the physical world manufacturing could rapidly compound for China. (01:15:37):
undefined

Dwarkesh: And also, I mean, their big advantage as a civilization and society, (01:15:41):
undefined

Dwarkesh: at least in recent decades, has been that they can do big industrial projects fast and efficiently. (01:15:44):
undefined

Dwarkesh: That's not the first thing you think of when you think of America. (01:15:50):
undefined

Dwarkesh: And AGI is a huge industrial, high CapEx, Manhattan project, right? (01:15:53):
undefined

Dwarkesh: And this is the kind of thing that China excels at and we don't. (01:16:00):
undefined

Dwarkesh: So, you know, I think it's like a much tougher race than people anticipate. (01:16:03):
undefined

Ryan Sean Adams: So what's all this going to do for the world? So once we get to the point of AGI, (01:16:07):
undefined

Ryan Sean Adams: we've talked about GDP and your estimate is less on the Tyler Cowen kind of (01:16:11):
undefined

Ryan Sean Adams: half a percent per year and more on, I guess, the Satya Nadella from Microsoft, (01:16:15):
undefined

Ryan Sean Adams: what does he say, 7% to 8% once we get to AGI. (01:16:21):
undefined

Ryan Sean Adams: What about unemployment? Does this cause mass, I guess, job loss across the (01:16:24):
undefined

Ryan Sean Adams: economy or do people adopt? (01:16:30):
undefined

Ryan Sean Adams: What's your take here? Yeah, what are you seeing? (01:16:33):
undefined

Dwarkesh: Yeah, I mean, definitely will cause job loss. I think people who don't, (01:16:36):
undefined

Dwarkesh: I think a lot of AI leaders try to gloss over that or something. And like, I mean. (01:16:39):
undefined

Josh Kale: What do you mean? (01:16:43):
undefined

Dwarkesh: Like, what does AGI mean if it doesn't cause job loss, right? (01:16:43):
undefined

Dwarkesh: If it does what a human does and. (01:16:45):
undefined

Josh Kale: It does it (01:16:47):
undefined

Dwarkesh: Cheaper and better and faster, like why would that not cause job loss? (01:16:48):
undefined

Dwarkesh: The positive vision here is just that it creates so much wealth, (01:16:52):
undefined

Dwarkesh: so much abundance, that we can still give people a much better standard of living (01:16:56):
undefined

Dwarkesh: than even the wealthiest people today, even if they themselves don't have a job. (01:17:00):
undefined

Dwarkesh: The future I worry about is one where instead of creating some sort of UBI that (01:17:06):
undefined

Dwarkesh: will get exponentially bigger as society gets wealthier, (01:17:12):
undefined

Dwarkesh: we try to create these sorts of guild-like protection rackets where if the coders got unemployed, (01:17:16):
undefined

Dwarkesh: then we're going to make these bullshit jobs just for the coders and this is (01:17:26):
undefined

Dwarkesh: how we give them a redistribution. (01:17:32):
undefined

Dwarkesh: Or we try to expand Medicaid for AI, but it's not allowed to procure all of (01:17:34):
undefined

Dwarkesh: these advanced medicines and cures that AI is coming up with, (01:17:42):
undefined

Dwarkesh: rather than just giving people, you know, maybe lump sums of money or something. (01:17:46):
undefined

Dwarkesh: So I am worried about the future where instead of sharing this abundance and (01:17:50):
undefined

Dwarkesh: just embracing it, we just have these protection rackets that maybe let a few (01:17:54):
undefined

Dwarkesh: people have access to the abundance of AI. (01:18:00):
undefined

Dwarkesh: So maybe like if you sue AI, if you sue the right company at the right time, (01:18:03):
undefined

Dwarkesh: you'll get a trillion dollars, but everybody else is stuck with nothing. (01:18:06):
undefined

Dwarkesh: I want to avoid that future and just be honest about what's coming and make (01:18:09):
undefined

Dwarkesh: programs that are simple and acknowledge how fast things will change and are (01:18:15):
undefined

Dwarkesh: forward looking rather than trying to turn what already exists into something (01:18:21):
undefined

Dwarkesh: amenable to the displacement that AI will create. (01:18:25):
undefined

Ryan Sean Adams: That argument reminds me of, I don't know if you read the essay recently came (01:18:29):
undefined

Ryan Sean Adams: out called The Intelligence Curse. Did you read that? (01:18:32):
undefined

Ryan Sean Adams: It was basically the idea of applying kind of the nation state resource curse (01:18:34):
undefined

Ryan Sean Adams: to the idea of intelligence. (01:18:40):
undefined

Ryan Sean Adams: So like nation states that are very high in natural resources, (01:18:42):
undefined

Ryan Sean Adams: they just have a propensity. (01:18:45):
undefined

Ryan Sean Adams: I mean, an example is kind of like a Middle Eastern state with lots of oil reserves, right? (01:18:48):
undefined

Ryan Sean Adams: They have this rich source of a commodity type of abundance. (01:18:53):
undefined

Ryan Sean Adams: They need their people less. And so they don't invest in citizens' rights. (01:18:58):
undefined

Ryan Sean Adams: They don't invest in social programs. (01:19:02):
undefined

Ryan Sean Adams: The authors of the intelligence curse were saying that there's a similar type (01:19:04):
undefined

Ryan Sean Adams: of curse that could happen once intelligence gets very cheap, (01:19:08):
undefined

Ryan Sean Adams: which is basically like the nation state doesn't need humans anymore. (01:19:11):
undefined

Ryan Sean Adams: And those at the top, the rich, wealthy corporations, they don't need workers anymore. (01:19:15):
undefined

Ryan Sean Adams: So we get kind of locked in this almost feudal state where, you know, (01:19:19):
undefined

Ryan Sean Adams: everyone has the property that their grandparents had and there's no meritocracy (01:19:23):
undefined

Ryan Sean Adams: and sort of the nation states don't reinvest in citizens. (01:19:27):
undefined

Ryan Sean Adams: Almost some similar ideas to your idea that like, you know, that the robots (01:19:31):
undefined

Ryan Sean Adams: might want us just, or sorry, the AIs might just want us for our meat hands (01:19:35):
undefined

Ryan Sean Adams: because they don't have the robotics technology on a temporary basis. (01:19:39):
undefined

Ryan Sean Adams: What do you think of this type of like future? Is this possible? (01:19:43):
undefined

Dwarkesh: I agree that that is like definitely more of a concern given that humans will (01:19:46):
undefined

Dwarkesh: not be directly involved in the economic output that will be generated in the CIA civilization. (01:19:49):
undefined

Dwarkesh: The hopeful story you can tell is that a lot of these Middle Eastern resource, (01:19:54):
undefined

Dwarkesh: you know, Dutch disease is another term that's used, (01:19:59):
undefined

Dwarkesh: countries, the problem is that they're not democracies, so that this wealth (01:20:01):
undefined

Dwarkesh: can just be, the system of government (01:20:06):
undefined

Dwarkesh: just lets whoever's in power extract that wealth for themselves. (01:20:08):
undefined

Dwarkesh: Whereas there are countries like Norway, for example, which also have abundant (01:20:11):
undefined

Dwarkesh: resources, who are able to use those resources to have further social welfare (01:20:15):
undefined

Dwarkesh: programs, to build sovereign wealth funds for their citizens, (01:20:21):
undefined

Dwarkesh: to invest in their future. (01:20:24):
undefined

Dwarkesh: We are going into, at least some countries, America included, (01:20:26):
undefined

Dwarkesh: will go into the age of AI as a democracy. (01:20:29):
undefined

Dwarkesh: And so we, of course, will lose our economic leverage, but the average person (01:20:33):
undefined

Dwarkesh: still has their political leverage. (01:20:38):
undefined

Dwarkesh: Now, over the long run, yeah, if we didn't do anything for a while, (01:20:39):
undefined

Dwarkesh: I'm guessing the political system would also change. (01:20:44):
undefined

Dwarkesh: So then the key is to lock in or turn our current, well, it's not just political leverage, right? (01:20:46):
undefined

Dwarkesh: We also have property rights. So like we own a lot of stuff that AI wants, factories, (01:20:52):
undefined

Dwarkesh: sources of data, et cetera, is to use the combination of political and economic (01:20:56):
undefined

Dwarkesh: leverage to lock in benefits for us for the long term, but beyond our the lifespan (01:21:01):
undefined

Dwarkesh: of our economic usefulness. (01:21:07):
undefined

Dwarkesh: And I'm more optimistic for us than I am for these Middle Eastern countries (01:21:10):
undefined

Dwarkesh: that started off poor and also with no democratic representation. (01:21:13):
undefined

Ryan Sean Adams: What do you think the future of like ChachipD is going to be? (01:21:17):
undefined

Ryan Sean Adams: If we just extrapolate maybe one version update forward to ChatGPT 5, (01:21:20):
undefined

Ryan Sean Adams: do you think the trend line of the scaling law will essentially hold for ChatGPT 5? (01:21:25):
undefined

Ryan Sean Adams: I mean, another way to ask that question is, do you feel like it'll feel like (01:21:30):
undefined

Ryan Sean Adams: the difference between maybe a BlackBerry and an iPhone? (01:21:33):
undefined

Ryan Sean Adams: Or will it feel more like the difference between, say, the iPhone 10 and the (01:21:37):
undefined

Ryan Sean Adams: iPhone 11, which is just like incremental progress, not a big breakthrough, (01:21:41):
undefined

Ryan Sean Adams: not an order of magnitude change? Yeah. (01:21:45):
undefined

Dwarkesh: I think it'll be somewhere in between but I don't think it'll feel like a humongous (01:21:50):
undefined

Dwarkesh: breakthrough even though I think it's in a remarkable pace of change because (01:21:53):
undefined

Dwarkesh: the nature of scaling is that sometimes people talk about it as an exponential process, (01:21:58):
undefined

Dwarkesh: Exponential usually refers to like it going like this. (01:22:03):
undefined

Dwarkesh: So having like a sort of J curve aspect to it, where the incremental input is (01:22:07):
undefined

Dwarkesh: leading to super linear amounts of output, in this case, intelligence and value, (01:22:10):
undefined

Dwarkesh: where it's actually more like a sideways J. (01:22:16):
undefined

Dwarkesh: The exponential means the exponential and the scaling laws is that you need (01:22:20):
undefined

Dwarkesh: exponentially more inputs to get marginal increases in usefulness or loss or intelligence. (01:22:23):
undefined

Dwarkesh: So and that's what we've been seeing, right? I think you initially see like some cool demo. (01:22:29):
undefined

Dwarkesh: So as you mentioned, you see some cool computer use demo, which comes at the (01:22:34):
undefined

Dwarkesh: beginning of this hyper exponential, I'm sorry, of this sort of plateauing curve. (01:22:38):
undefined

Dwarkesh: And then it's still an incredibly powerful curve and we're still early in it. (01:22:44):
undefined

Dwarkesh: But the next demo will be just adding on to making this existing capability (01:22:48):
undefined

Dwarkesh: more reliable, applicable for more skills. (01:22:54):
undefined

Dwarkesh: The other interesting incentive in this industry is that because there's so (01:22:57):
undefined

Dwarkesh: much competition between the labs, you are incentivized to release a capability. (01:23:01):
undefined

Dwarkesh: As soon as it's even marginally viable or marginally cool so you can raise more (01:23:06):
undefined

Dwarkesh: funding or make more money off of it. (01:23:11):
undefined

Dwarkesh: You're not incentivized to just like sit on it until you perfected it, (01:23:13):
undefined

Dwarkesh: which is why I don't expect like tomorrow OpenAI will just come out with like, (01:23:17):
undefined

Dwarkesh: we've solved continual learning, guys, and we didn't tell you about it. (01:23:20):
undefined

Dwarkesh: We're working on it for five years. (01:23:22):
undefined

Dwarkesh: If they had like even an inkling of a solution, they'd want to release it ASAP (01:23:24):
undefined

Dwarkesh: so they can raise a $600 billion round and then spend more money on compute. (01:23:27):
undefined

Dwarkesh: So yeah, I do think it'll seem marginal. But again, marginal in the context of seven years to AGI. (01:23:32):
undefined

Dwarkesh: So zoom out long enough and a crazy amount of progress is happening. (01:23:38):
undefined

Dwarkesh: Month to month, I think people overhype how significant any one new release is. So I guess the answer. (01:23:42):
undefined

Dwarkesh Patel: To when we will get AGI very much depends on that scaling trend holding. (01:23:48):
undefined

Dwarkesh Patel: Your estimate in the book for AGI was 60% chance by 2040. (01:23:52):
undefined

Dwarkesh Patel: So I'm curious, what guess or what idea had the most influence on this estimate? (01:23:57):
undefined

Dwarkesh Patel: What made you end up on 60% of 2040? (01:24:00):
undefined

Dwarkesh Patel: Because a lot of timelines are much faster than that. (01:24:04):
undefined

Dwarkesh: It's sort of reasoning about the things they currently still lack, (01:24:06):
undefined

Dwarkesh: the capabilities they still lack, and what stands in the way. (01:24:10):
undefined

Dwarkesh: And just generally an intuition that things often take longer to happen than (01:24:13):
undefined

Dwarkesh: you might think. Progress tends to slow down. (01:24:16):
undefined

Dwarkesh: Also, it's the case that, look, you might have heard the phrase that we keep (01:24:19):
undefined

Dwarkesh: shifting the goalposts on AI, right? (01:24:23):
undefined

Dwarkesh: So they can do the things which skeptics were saying they couldn't ever do already. (01:24:26):
undefined

Dwarkesh: But now they say AI is still a dead end because problem X, Y, (01:24:30):
undefined

Dwarkesh: Z, which will be solved next year. (01:24:34):
undefined

Dwarkesh: Now, there's a way in which this is frustrating, but there's another way in which there's some, (01:24:36):
undefined

Dwarkesh: It is the case that we didn't get to AGI, even though we have passed the Turing (01:24:43):
undefined

Dwarkesh: test and we have models that are incredibly smart and can reason. (01:24:46):
undefined

Dwarkesh: So it is accurate to say that, oh, we were wrong and there is some missing thing (01:24:49):
undefined

Dwarkesh: that we need to keep identifying about what is still lacking to the path of AGI. (01:24:53):
undefined

Dwarkesh: Like it does make sense to shift the goalposts. And I think we might discover (01:24:57):
undefined

Dwarkesh: once continual learning is solved or once extended computer use is solved, (01:25:00):
undefined

Dwarkesh: that there were other aspects of human intelligence, which we take for granted (01:25:04):
undefined

Dwarkesh: in this Moravax paradox sense, but which are actually quite crucial to making (01:25:08):
undefined

Dwarkesh: us economically valuable. (01:25:12):
undefined

Ryan Sean Adams: Part of the reason we wanted to do this, Dwarkesh, is because we both are enjoyers (01:25:14):
undefined

Ryan Sean Adams: of your podcast. It's just fantastic. (01:25:18):
undefined

Ryan Sean Adams: And you talk to all of the, you know, those that are on the forefront of AI (01:25:20):
undefined

Ryan Sean Adams: development, leading it in all sorts of ways. (01:25:25):
undefined

Ryan Sean Adams: And one of the things I wanted to do with reading your book, (01:25:28):
undefined

Ryan Sean Adams: and obviously I'm always asking myself when I'm listening to your podcast is (01:25:30):
undefined

Ryan Sean Adams: like, what does Dwarkesh think personally? (01:25:34):
undefined

Ryan Sean Adams: And I feel like I sort of got that insight maybe toward the end of your book, (01:25:36):
undefined

Ryan Sean Adams: like, you know, in the summary section, where you think like there's a 60% probability (01:25:39):
undefined

Ryan Sean Adams: of AGI by 2040, which puts you more in the moderate camp, right? (01:25:44):
undefined

Ryan Sean Adams: You're not a conservative, but you're not like an accelerationist. (01:25:48):
undefined

Ryan Sean Adams: So you're moderate there. (01:25:50):
undefined

Ryan Sean Adams: And you also said you think more than likely AI will be net beneficial to humanity. (01:25:51):
undefined

Ryan Sean Adams: So you're more optimist than Doomer. So we've got a moderate optimist. (01:25:57):
undefined

Ryan Sean Adams: And you also think this, and this is very interesting, There's no going back. (01:26:01):
undefined

Ryan Sean Adams: So you're somewhat of an AI determinist. And I think the reason you state for (01:26:05):
undefined

Ryan Sean Adams: not, you're like, there's no going back. (01:26:10):
undefined

Ryan Sean Adams: It struck me, there's this line in your book. It seems that the universe is (01:26:12):
undefined

Ryan Sean Adams: structured such that throwing large amounts of compute at the right distribution of data gets you AI. (01:26:16):
undefined

Ryan Sean Adams: And the secret is out. If the scaling picture is roughly correct, (01:26:21):
undefined

Ryan Sean Adams: it's hard to imagine AGI not being developed this century, even if some actors (01:26:24):
undefined

Ryan Sean Adams: hold back or are held back. (01:26:28):
undefined

Ryan Sean Adams: That to me is an AI determinist position. Do you think that's fair? (01:26:31):
undefined

Ryan Sean Adams: Moderate with respect to accelerationism, optimistic with respect to its potential, (01:26:34):
undefined

Ryan Sean Adams: and also determinist, like there's nothing else we can do. We can't go backwards here. (01:26:39):
undefined

Dwarkesh: I'm determinist in the sense that I think if AI is technologically possible, it is inevitable. (01:26:43):
undefined

Dwarkesh: I think sometimes people are optimistic about this idea that we as a world will sort of, (01:26:48):
undefined

Dwarkesh: I collectively decide not to build AI. And I just don't think that's a plausible outcome. (01:26:53):
undefined

Dwarkesh: The local incentives for any actor to build AI are so high that it will happen. (01:26:58):
undefined

Dwarkesh: But I'm also an optimist in the sense that, look, I'm not naive. (01:27:02):
undefined

Dwarkesh: I've listed out all the way, like what happened to the Aztecs and Incas was (01:27:05):
undefined

Dwarkesh: terrible. And I've explained how that could be similar to what AIs could do (01:27:08):
undefined

Dwarkesh: to us and what we need to do to avoid that outcome. (01:27:10):
undefined

Dwarkesh: But I am optimistic in the sense that the world of the future fundamentally (01:27:13):
undefined

Dwarkesh: will have so much abundance that there's all these, (01:27:18):
undefined

Dwarkesh: that alone is a prima facie reason to think that there must be some way of cooperating (01:27:22):
undefined

Dwarkesh: that is mutually beneficial. (01:27:28):
undefined

Dwarkesh: If we're going to be thousands, millions of times wealthier, (01:27:30):
undefined

Dwarkesh: is there really no way that humans are better off or can we can find a way for (01:27:33):
undefined

Dwarkesh: humans to become better off as a result of this transformation? (01:27:37):
undefined

Dwarkesh: So yeah, I think you've put your finger on it. (01:27:40):
undefined

Ryan Sean Adams: So this scaling book, of course, goes through the history of AI scaling. (01:27:43):
undefined

Ryan Sean Adams: I think everyone should should pick it up to get the full chronology, (01:27:46):
undefined

Ryan Sean Adams: but also sort of captures where we are in the midst of this story is like, we're not done yet. (01:27:49):
undefined

Ryan Sean Adams: And I'm wondering how you feel at this moment of time. (01:27:55):
undefined

Ryan Sean Adams: So I don't know if we're halfway through, if we're a quarter way through, (01:27:58):
undefined

Ryan Sean Adams: if we're one tenth of the way through, but we're certainly not finished the path to AI scaling. (01:28:03):
undefined

Ryan Sean Adams: How do you feel like in this moment in 2025? (01:28:07):
undefined

Ryan Sean Adams: I mean, is all of this terrifying? Is it exciting? (01:28:11):
undefined

Ryan Sean Adams: Is it exhilarating? (01:28:15):
undefined

Ryan Sean Adams: What's the emotion that you feel? (01:28:17):
undefined

Dwarkesh: Maybe I feel a little sort of hurried. I personally feel like there's a lot (01:28:20):
undefined

Dwarkesh: of things I want to do in the meantime, (01:28:24):
undefined

Dwarkesh: including what my mission is with the podcast, which is to, and I know it's (01:28:27):
undefined

Dwarkesh: your mission as well, is to improve the discourse around these topics, (01:28:33):
undefined

Dwarkesh: to not necessarily push for a specific agenda, but make sure that when people are making decisions, (01:28:38):
undefined

Dwarkesh: they're as well-informed as possible, They have as much strategic awareness (01:28:42):
undefined

Dwarkesh: and depth of understanding around how AI works, what it could do in the future as possible. (01:28:47):
undefined

Dwarkesh: And, but in many ways, I feel like I still haven't emotionally priced in the future I'm expecting. (01:28:55):
undefined

Dwarkesh: In this one very basic sense, I think that there's a very good chance that I (01:29:00):
undefined

Dwarkesh: live beyond 200 years of age. (01:29:06):
undefined

Dwarkesh: I have not changed anything about my life with regards to that knowledge, right? (01:29:08):
undefined

Dwarkesh: I'm not like, when I'm picking partners, I'm not like, oh, this is the person, (01:29:12):
undefined

Dwarkesh: now that I think I'm going to live for 200, you know, like hundreds of years. (01:29:17):
undefined

Ryan Sean Adams: Yeah. (01:29:20):
undefined

Dwarkesh: Well, you know, ideally I would pick a partner that would, ideally you pick (01:29:23):
undefined

Dwarkesh: somebody who would be, that would be true regardless. (01:29:27):
undefined

Dwarkesh: But you see what I'm saying, right? There's like, the fact that I expect my (01:29:30):
undefined

Dwarkesh: personal life, the world around me, the lives of the people I care about, (01:29:34):
undefined

Dwarkesh: humanity in general to be so different has, it just like doesn't emotionally resonate as much as, (01:29:37):
undefined

Dwarkesh: I, my intellectual thoughts and my emotional landscape aren't in the same place. (01:29:45):
undefined

Dwarkesh: I wonder if it's similar for you guys. (01:29:50):
undefined

Ryan Sean Adams: Yeah, I totally agree. I don't think I've priced that in. Also, (01:29:51):
undefined

Ryan Sean Adams: there's like non-zero chance that Eliezer Yudkowsky is right, Dworkesh. (01:29:54):
undefined

Ryan Sean Adams: Do you know? And so that scenario, I just, I can't bring myself to emotionally price in. (01:29:58):
undefined

Ryan Sean Adams: So I veer towards the optimism side as well. (01:30:03):
undefined

Ryan Sean Adams: Dworkesh, this has been fantastic. Thank you so much for all you do on the podcast. (01:30:07):
undefined

Ryan Sean Adams: I have to ask a question for our crypto audience as well, which is, (01:30:12):
undefined

Ryan Sean Adams: when are you going to do a crypto podcast on Dwarkech? (01:30:15):
undefined

Dwarkesh: I already did. It was with one Sam Bigman-Fried. (01:30:19):
undefined

Ryan Sean Adams: Oh my God. (01:30:22):
undefined

Dwarkesh: Oh man. (01:30:24):
undefined

Ryan Sean Adams: We got to get you a new guest. We got to get you someone else to revisit the top best. (01:30:26):
undefined

Dwarkesh: Don't look that one up. It's Ben Omen. Don't look that one up. (01:30:29):
undefined

Dwarkesh: I think in retrospect. You know what? We'll do another one. (01:30:31):
undefined

Ryan Sean Adams: Fantastic. I'll ask you (01:30:36):
undefined

Dwarkesh: Guys for some recommendations. That'd be great. Dwarkech, thank you so much. (01:30:37):
undefined

Dwarkesh: But I've been following your stuff for a while, for I think many years. (01:30:40):
undefined

Dwarkesh: So it's great to finally meet. and this was a lot of fun. (01:30:43):
undefined

Ryan Sean Adams: Appreciate it. It was great. Thanks a lot. (01:30:46):
undefined
Advertise With Us

Popular Podcasts

Crime Junkie

Crime Junkie

Does hearing about a true crime case always leave you scouring the internet for the truth behind the story? Dive into your next mystery with Crime Junkie. Every Monday, join your host Ashley Flowers as she unravels all the details of infamous and underreported true crime cases with her best friend Brit Prawat. From cold cases to missing persons and heroes in our community who seek justice, Crime Junkie is your destination for theories and stories you won’t hear anywhere else. Whether you're a seasoned true crime enthusiast or new to the genre, you'll find yourself on the edge of your seat awaiting a new episode every Monday. If you can never get enough true crime... Congratulations, you’ve found your people. Follow to join a community of Crime Junkies! Crime Junkie is presented by audiochuck Media Company.

24/7 News: The Latest

24/7 News: The Latest

The latest news in 4 minutes updated every hour, every day.

Stuff You Should Know

Stuff You Should Know

If you've ever wanted to know about champagne, satanism, the Stonewall Uprising, chaos theory, LSD, El Nino, true crime and Rosa Parks, then look no further. Josh and Chuck have you covered.

Music, radio and podcasts, all free. Listen online or download the iHeart App.

Connect

© 2025 iHeartMedia, Inc.