<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Dwarkesh Podcast: Blog]]></title><description><![CDATA[All my posts]]></description><link>https://www.dwarkesh.com/s/blog</link><image><url>https://substackcdn.com/image/fetch/$s_!QEPJ!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F90fa9666-5b8b-4685-a8fb-4b64cb7e0333_1080x1080.png</url><title>Dwarkesh Podcast: Blog</title><link>https://www.dwarkesh.com/s/blog</link></image><generator>Substack</generator><lastBuildDate>Tue, 28 Apr 2026 11:46:07 GMT</lastBuildDate><atom:link href="https://www.dwarkesh.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Dwarkesh Patel]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[dwarkesh@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[dwarkesh@substack.com]]></itunes:email><itunes:name><![CDATA[Dwarkesh Patel]]></itunes:name></itunes:owner><itunes:author><![CDATA[Dwarkesh Patel]]></itunes:author><googleplay:owner><![CDATA[dwarkesh@substack.com]]></googleplay:owner><googleplay:email><![CDATA[dwarkesh@substack.com]]></googleplay:email><googleplay:author><![CDATA[Dwarkesh Patel]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[What I've been thinking about this weekend - More open questions, intelligence vs power, the problem of verification in science, the parallel discovery of Darwinism]]></title><description><![CDATA[Hodge podge of things I was thinking about this weekend.]]></description><link>https://www.dwarkesh.com/p/what-ive-been-thinking-april-27</link><guid isPermaLink="false">https://www.dwarkesh.com/p/what-ive-been-thinking-april-27</guid><dc:creator><![CDATA[Dwarkesh Patel]]></dc:creator><pubDate>Mon, 27 Apr 2026 13:51:32 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/7b074c57-7c3d-48b6-a304-2cea506b5165_800x450.avif" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>More open questions</h2><p>I put out <a href="https://www.dwarkesh.com/p/blog-prize">a blog prize</a> to answer a couple of big questions I have about AI. The goal is really to find someone to hire as coresearcher. I have more questions of this variety, but I omitted them from that post&#8217;s list, because they don&#8217;t make it easy to judge submission quality. So I thought I&#8217;d post them here:</p><ul><li><p>5 hyperscalers own 70+% of global AI compute, and much of that is actually reserved for the 3 member set of OpenAI/Ant/GDM. How worried should we be that AI use cases which are not building up to the singularity and the robot factories -  aka normal people being more empowered, understanding the world better, being entertained, etc, is not the highest ROI activity for compute in the world. And given how valuable compute will be (whose opportunity cost increases in tandem with the quality of the AI models that run on it), will normal people basically get priced out of the benefits of AI? If we should be worried about this, how concretely should some kind of universal basic income/compute redistribution work? If not worried, what is the frame of this question missing?</p></li><li><p>Data is arguably the main way that AI models have been getting better over the last few years. But I remain confused about what concretely these improvements have consisted of. To ask some sharper questions:</p><ul><li><p>Clearly Anthropic (and now also OpenAI and GDM) have cracked <em>something</em> about making competent long horizon coding agents. What is it? Is it just stacking up more and more RL coding environments? Or is there something more particular behind this breakthrough?</p></li><li><p>Are models even getting more sample efficient (aka they learn more from each training sample) or have we just changed/expanded/improved the data input? The reason this question is important is because it tells us how fast deep learning progress will be in domains that actually do require sample efficiency (for example, robotics).</p><ul><li><p>Models are very sample efficient in context, and the information in context can be used much more flexibly. But the attention &#8220;fast&#8221; weights consume a huge amount of memory in order to accommodate this faster learning. Why is there this memory/sample efficiency tradeoff?</p><ul><li><p>If you look at the size of the KV cache for Llama 3 70B, it&#8217;s 320 KB / token. If you just divide the number of bits it takes to store Llama 3 weights by the number of tokens it was pre-trained on, then you get 0.075 bits / token. So there&#8217;s a 35 million fold difference in the amount of information per bit you&#8217;re storing.</p></li></ul></li></ul></li></ul></li><li><p>Let&#8217;s put frontier lab compute into 3 buckets: pretraining, RL generation, and inference. RL generation and inference look like very similar workloads. The big difference, of course, is that the model learns as a result of RL generation, but it doesn&#8217;t (at least currently) from inference. At the same time, the model actually does useful work during inference, but not during RL generation. Many people have pointed out it&#8217;s really weird that there&#8217;s a distinction between training and inference, and that in the limit it shouldn&#8217;t exist. How practically will these two workloads be merged? At a high level, one can imagine hiring an AI instance for a month-long work trial, getting it to do actual useful work for you during that time, and then sending a report card back to the model company. In fact, in a few years, maybe the only way that AI can continue to make progress is through this kind of on-the-job learning, because models will already have saturated anything that can be learned from contrived shorter-horizon RL environments.</p></li><li><p>Does something Y2Key happen when most of the tokens on the internet (and presumably the ones future models will be trained on) are generated by other AIs? Has the relative value of pre-2023 internet datasets increased in any noticeable way?</p></li><li><p>I wrote this in my <a href="http://dwarkesh.com/p/timelines-june-2025">continual learning blog post</a> last June. Is this correct? Why might there not be a winner take all dynamic from continual learning?</p><ul><li><p>&#8220;Even if there isn&#8217;t a software only singularity (with models rapidly building smarter and smarter successor systems), we might still see something that looks like a broadly deployed intelligence explosion. AIs will be getting broadly deployed through the economy, doing different jobs and learning while doing them in the way humans can. But unlike humans, these models can amalgamate their learnings across all their copies. So one AI is basically learning how to do every single job in the world. An AI that is capable of online learning might functionally become a superintelligence quite rapidly without any further algorithmic progress&#8221;</p></li></ul></li><li><p>A lot of economic analysis about the impact of AGI focus on human demand - <a href="https://www.citriniresearch.com/p/2028gic">will the economy shrink because our demands can be fulfilled much more cheaply</a>, will it rise because <a href="https://aleximas.substack.com/p/what-will-be-scarce/comment/243687560">AI will create new varieties of products</a>, or maybe because <a href="https://aleximas.substack.com/p/what-will-be-scarce">the relational sector will grow</a>? But all this analyses take as a given that the only demand that matters is the one originating from humans. How do we model the machine-only economy, where the demand originates from the AI&#8217;s themselves? And once we add this consideration to our economic analysis of the future, what changes?</p></li></ul><h2>The mistake of conflating intelligence and power</h2><p>I had an interesting discussion recently. Someone asked me, what is intelligence? I said, the ability to achieve your goals across a wide range of domains. Okay, he says, then by that definition isn&#8217;t Donald Trump the intelligent person in the person, followed quickly by Xi Jinping and Vladimir Putin?</p><p>To be clear, these people are obviously very competent in certain ways. But when you think of ASI, you don&#8217;t think of Trump, but more so. The person who kept pressing this question was correctly pointing out that my definition of intelligence was basically power (after all, what is power if not the ability to achieve your goals across a wide range of domains?). If this is your definition of intelligence, then Stalin was the most intelligent person who ever lived.</p><p>Now, of course, you could change the definition of intelligence to something more like, comprehend and build atop abstract concepts. But notice that the most powerful people in the world do not max out this quantity. They&#8217;re above average in shape-rotation, but the correlation between extreme power and this kind of intelligence might be even weaker than the correlation between extreme power and height. The physicists are not running the world.</p><p>We tend to conflate power-seeking AI and superintelligent (in science and tech) AI. I&#8217;m not denying that AI can be power-seeking. Whatever skills and drives Donald Trump has could be embodied in a digital mind. I&#8217;m simply pointing out that the way we&#8217;re currently making AI systems smarter (training them to be really good coders, thought partners, and general coworkers) is not that strongly correlated with power.</p><p>We often talk about power in this way that misunderstands how it is actually derived in our world. Our intuitions are primed by games like Diplomacy or Go, which are designed to isolate and reward a g loaded kind of strategic reasoning. But in the real world, power is more the product of having the authority and trust to get lots of people to collaborate with you, rather than some galaxy brain scheming capability. Trump is not powerful because his brain, considered in isolation, is the most effective optimization engine on Earth. He is powerful because the government which hundreds of millions of people consider legitimate gives him a lot of power.</p><p>A group versus individual level analysis is useful here. As <a href="https://mason.gmu.edu/~gjonesb/JonesADR">Garett Jones has written</a> a lot about, individual IQ is only modestly correlated with individual income, but national IQ is strongly correlated with national outcomes. This is because intelligence has a lot of spillover effects - smarter societies cooperate more, save more, and can coordinate to build things like space shuttles and semiconductors. Richard Trevithick, who pioneered the high-pressure steam engine, died in poverty, buried in an unmarked pauper&#8217;s grave. But the fact that 18th and 19th century Britain had lots and lots of people like Trevithick contributed to Britain being able to set up a global empire and defeat lots of random kings and emperors around the world. George III himself didn&#8217;t need to be a genius &#8212; in fact he went mad halfway through his reign &#8212; but the country he sat atop still defeated Napoleon, conquered India, and built the world&#8217;s dominant navy. Similarly, even if some company&#8217;s AIs are just super obedient superintelligent coders and scientists, they could help the totally pedestrian human intelligences who have their reins (lab leaders, Presidents, some harder to imagine configuration of control) gain a lot of power. It seems to me that the right mental model is that more effective AI firms and countries will outcompete everyone else in normal capitalist ways, rather than a single AI outthinking everyone else.</p><h2>RLVR might be disproportionately <em>bad</em> at science</h2><p><em>Next two sections I&#8217;m writing up some threads that we explored in <a href="https://www.dwarkesh.com/p/michael-nielsen">my interview with Michael Nielson</a>. That episode was one of my favorite.</em></p><p>The organizing question from my interview with Nielson was, &#8220;How do we recognize scientific progress?&#8221; It&#8217;s especially relevant to thinking about what it would take for AI to close the RL verification loop on scientific discovery. But it&#8217;s also a surprisingly mysterious and elusive question when thinking about the history of human science.</p><p>Some people have this idea that AI is going to be disproportionately good at making scientific breakthroughs. The reason they think this is that 1. Science is &#8216;verifiable&#8217;, 2. AI is absolutely crushing domains that have a tight verification loop - coding, math, etc - because you can RL on these loops.</p><p>But the history of human science shows that the verification loop for theories can be on the order of decades and centuries, and even then experiments do not definitely rule out alternatives: Ancient Athenians dismissed Aristarchus (2nd century BC) on heliocentrism because it would imply stellar parallax. The first successful measurement of stellar parallax was in 1838, achieved by Friedrich Wilhelm Bessel.</p><p>What we know today as the better theory can often actually make worse predictions: it&#8217;s well known that Copernicus&#8217;s model of circular orbits around the sun was less accurate than Ptolemy&#8217;s geocentric model, which had accumulated millenia of correcting epicycles. What is not well known is that Copernicus&#8217;s theory wasn&#8217;t even simpler (Ptolemy&#8217;s model interpreted the true elliptical nature of orbits using an equant trick where other planets are not moving in uniform circular motion around Earth exactly, but rather an off center point. Copernicus didn&#8217;t like this, because it violated his Platonic heuristics  - so he discarded the quant trick, which led to a less parsimonious model, since Copernicus had to add more epicycles and epicyclets to make up for it.)</p><p>So in what sense was it a better theory in 1543? In some sense, it wasn&#8217;t! You couldn&#8217;t have known ex ante that heliocentrism married with Kepler&#8217;s 3 laws (1619) is a much cleaner and more accurate theory, or that there&#8217;s a very beautiful unification of heliocentric orbits and terrestrial gravity (Newton in 1686).</p><p>There was one ex ante reason that you should have preferred Copernicus in 1543: his theory required retrograde motion<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> as a natural consequence of his theory, whereas for Ptolemy it was an ad hoc addition. Even more impressively, his theory, developed in 1543, actually predicted the phases of Venus<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> before they were observed by Galileo in 1610. But both of these things were also implied by Brahe&#8217;s model, which had set the sun to orbit the earth and then all the planets to orbit the sun.</p><p>Under a naive falsificationist framework, you&#8217;d have to wait until Stellar parallax was observed in 1838 to know that Brahe was wrong. But obviously the scientific community was able to make progress faster than this. There is some mixture of judgment and heuristics in the progress of science that we don&#8217;t even understand well enough to actually articulate, much less codify into an RL loop.</p><p>Or consider the case of the discovery of Neptune in 1846. Uranus deviated from its predicted Newtonian path. Le Verrier predicted that an unknown perturbing planet must exist, calculated its mass and orbit, and Neptune was found almost exactly where predicted.</p><p>But the Neptune story is symmetric to a failure case. Mercury had an anomalous precession, where the ellipse that shows its orbit would rotate 43 arcseconds more per century than should be implied by the impact of other planets using Newtonian mechanics. This led astronomers to speculate that there&#8217;s an unknown planet Vulcan within Mercury&#8217;s orbit. But it was resolved in 1915 with Einstein&#8217;s General Relativity.</p><p>A proper Newtonian would still proceed with the research agenda, but modify it as follows. First, you predict some unknown planet. If it can&#8217;t be found, you say it&#8217;s so small, it must require a bigger telescope, and you build a bigger telescope. And if you still can&#8217;t find it, maybe there&#8217;s a cloud of cosmic dust occluding it. If still not found, maybe the satellite&#8217;s instruments are being screwed by some unknown magnetic field, and you send a new satellite. At each of these steps, had you discovered a new planet, or some unknown cosmic dust, or some new magnetic field, that would have been a sensational victory for Newtonians.</p><p>Ex ante, this is not unreasonable to do! It is only after decades or maybe centuries of patchwork that we can then analyze, are we simply adding epicycles, or is this theoretical framework progressive, in that it makes predictions we wouldn&#8217;t otherwise be able to.</p><p>What do these examples illustrate? That ex ante it is almost impossible to determine which research programs are progressive (will predict and explain unanticipated new phenomenon) and which are regressive (need to be contorted repeatedly to accommodate seemingly disconfirming new phenomenon).</p><p>But the verification loop is often extremely long and weirdly hostile, and even then, experiments do not definitely rule out alternatives (see the discussion in the Nielson episode about how physicist contemporaneous with the 1880s Michelson-Morley experiments thought that it simply ruled out a particular theory of ether. Only Einstein made the full conceptual leap to discard the ether altogether).</p><p>This means that big conceptual breakthroughs <em>cannot</em> be easily verified. They are recognized decades  or centuries later, when it turns out they were much more productive than the alternatives available. What this means for AI for science is that 1. You can&#8217;t easily train an RL loop for big conceptual breakthroughs.</p><p>And 2. the society of AI scientists will still need individual AI instances that have idiosyncratic biases and heuristics, and to pursue them unrelentingly for decades on end - for example, like the one Einstein had in insisting that there shouldn&#8217;t be some arbitrary inertial reference frame. There should be dedicated people to keep a bunch of dormant research agendas alive in case they turn out to be productive upon further investigation. To understand the kind of intransigent dedication to hypotheses that is needed to preserve correct scientific idea - even in the face of disconfirming evidence - consider the following story: In 1815, Prout hypothesized that the atomic weights of all pure chemical elements are whole numbers, because experimentally, most elements seem to come out like this. But there&#8217;s many anomalies - for example Chlorine&#8217;s atomic weight is measured at 35.5. And so Prout&#8217;s school claimed that maybe the chemical substance in which these elements appeared were impure. But there seemed to be no chemical reaction that could get rid of the impurities. And then they said, maybe it&#8217;s fractions of full atomic weights - but the closer you measure, the less natural the fractions seem to get - Chlorine goes from 35.5 to 35.46. It takes until almost a century later for people to realize that these measurements are showing multiple isotopes of the same element, which can be separated physically, but have no chemical distinguishing characteristics.</p><p>What I&#8217;m trying to say is that ex ante, one couldn&#8217;t have known which research program would be more productive. We need to invest in all of them concurrently. But that investment looks like a bunch of different individual scientists being super unreasonable and obstinate about propping up their preferred research agenda.</p><h2>What does the parallel discovery of a deep idea like Darwinism tell us?</h2><p>The Origin of Species was published in 1859. Principia Mathematica was published in 1687, two centuries earlier. Conceptually, it seems like natural selection is much simpler than the theory of gravity. A contemporary of Darwin&#8217;s, Thomas Huxley, read the Origin of Species and said, &#8220;How extremely stupid not to have thought of that!&#8221; Nobody ever said the same for not beating Newton to the Principia. I wonder if the reason this happened is that, while Darwin&#8217;s theory is conceptually simpler, it cannot be decisively tested. The evidence is circumstantial, retrospective, and cumulative. There&#8217;s no equivalent of Newton running the numbers on the moon&#8217;s orbital period and radius, and confirming that it corresponds to his equations.</p><p>Also you need this concept of deep time. Charles Lyell published the Principles of Geology in 1830, which gave Darwin the vast stretches of time that natural selection needed. And the fact that Darwin and Wallace basically arrived at evolution at the same time<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> (and both credited Lyell&#8217;s contribution) does suggest that these underrated intellectual footholds were quite important (geology, paleontology of ancient extinct species which showed intermediate species (in some cases between apes and humans), biogeography from voyages and age of colonization, more sophisticated artificial selection like pigeon breeding). It&#8217;s interesting that an idea whose essence must have been obvious to herders and parents for thousands of years actually required many millennia of ancillary intuition pumps to fully spell out.</p><p>The pattern of parallel discovery in science and technology is very interesting, and seems to contradict this vibe that certain innovations could have happened earlier much earlier than they really did.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>where Mars appears to slow down and reverse direction as Earth overtakes it with its faster inner orbit</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>since Venus&#8217;s orbit is inside Earth&#8217;s, you should see if fully dark when it&#8217;s between Earth and Sun, and crescent halfway through, and fully lit up when it&#8217;s on the other side, aka when it&#8217;s smallest</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>their work was shown as a joint presentation at the Linnean Society in 1858.</p></div></div>]]></content:encoded></item><item><title><![CDATA[Blog prize for the big questions about AI]]></title><description><![CDATA[The not-so-secret point of this whole contest is so that I can hire a researcher]]></description><link>https://www.dwarkesh.com/p/blog-prize</link><guid isPermaLink="false">https://www.dwarkesh.com/p/blog-prize</guid><dc:creator><![CDATA[Dwarkesh Patel]]></dc:creator><pubDate>Fri, 24 Apr 2026 16:37:49 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/a7d14f96-c3aa-4305-bdc2-27c509bdbedc_1400x923.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>There has never been a time where excellent intellectual output on the right question has been more valuable or more urgent. Compelling answers can inform the most important economic and foreign policy decisions that will ever be made, the deployment of (at least) <a href="https://openai.com/index/scaling-ai-for-everyone/">hundreds of billions</a> of philanthropic dollars, and the training and governance of superintelligences.</p><p>I&#8217;m announcing a $20,000 blog prize in order to find people who will excel at researching and thinking through these problems. The not-so-secret point of this whole contest is so that I can hire a research collaborator to think through questions like this hand in hand with me. See more at the end.</p><p>Pick a question below, and spend no more than 1,000 words answering it. 1st, 2nd, and 3rd place will get $10,000, $6,000, and $4,000 respectively. I&#8217;ll publish the winning entry (and potentially the runner ups) on my blog. Please submit by May 10th, 11:59 PM PST.</p><h3>Questions - choose one</h3><ul><li><p>A couple years ago, there was this idea that AI progress might slow down as we make further progress into the RL regime. 1. Because as horizon lengths increase, the AI needs to do many days&#8217; worth of work before we can even see if it did it right, so if we&#8217;re still in a naive policy gradient world, the reward signal / FLOP goes down, and 2. We&#8217;d crossed through many OOMs of RL compute from GPT 4 to o1 to o3, and it would not be feasible to replicate that many OOMs increase in compute immediately again. But AI progress seems to have been fast nonetheless - even potentially speeding up if rumors about Spud or Mythos are to be believed. What gives? What did that previous intuition pump that motivated longer timelines miss? Feel free to deny premise of question.</p></li><li><p>What&#8217;s the most plausible story where foundation model companies actually start making money? If you consider each individual model as a company, then its profits <a href="https://epoch.ai/gradient-updates/can-ai-companies-become-profitable">may</a> be able to pay back the training cost. But of course, if you don&#8217;t train a bigger, more expensive model immediately, then you stop making money after 3 months. So when does the profit start? Maybe at some point <a href="https://www.dwarkesh.com/i/187852154/005849-how-will-ai-labs-actually-make-profit">scaling will plateau</a>, but <a href="https://x.com/MatthewJBar/status/2046060153678844290">if progress at the frontier</a> has slowed down, then the combination of distillation and low switching costs (cloud margins result from high switching costs) makes it really easy for open source to catch up to the labs, eating into their margins. So how do the labs actually start making money?</p></li><li><p>With OpenAI&#8217;s new raise at an $852B valuation, OpenAI Foundation&#8217;s stake is <a href="https://openai.com/index/scaling-ai-for-everyone/">now worth $180B</a>. Anthropic&#8217;s <a href="https://fortune.com/2026/01/27/anthropic-billionaire-cofounders-ceo-dario-amodei-giving-away-80-percent-of-wealth-fighting-inequality-ai-revolution/">cofounders have pledged to donate 80%</a> of their wealth. Nobody seems to have a concrete idea of how to deploy 100s of billions (soon trillions) of wealth productively to &#8220;make AI go well&#8221;. If you were in charge of the OpenAI Foundation right now, what exactly would you do? And when? It&#8217;s not enough to identify a cause you think is important, because that doesn&#8217;t answer the fundamental problem of <a href="https://nanransohoff.substack.com/p/there-should-be-general-managers">how you convert money to impact</a>. Identify the concrete strategy you recommend pursuing.</p></li><li><p>What should countries which are not currently in the AI production chain (semis, energy, frontier models, robotics) do in order to not get totally sidestepped by transformative AI? If you&#8217;re the leader of India or Nigeria, what do you do right now?</p></li></ul><h3>Rules and tips</h3><ul><li><p>Please don&#8217;t let a lack of domain expertise dissuade you from entering. I&#8217;m looking for someone who can ramp up fast on unfamiliar topics and think clearly.</p></li><li><p>Each entrant may submit only once.</p></li><li><p>You are still eligible for this essay competition even if you&#8217;re not interested in the researcher role. Nor does winning this competition guarantee that you will be offered the role.</p></li><li><p>You&#8217;re welcome to use LLMs to help you research, but I specifically picked these questions because I&#8217;ve found LLM answers to them unsatisfying. On these kinds of ambiguous questions, LLMs are too all over the place. For example, they&#8217;ll identify 5 plausible answers but not have the context and taste to identify the crucial factor and iron out its implications.</p></li><li><p>You only have 1000 words - make them count. People have the habit of <a href="https://x.com/dwarkesh_sp/status/1968012981016608934">spending the first paragraphs clearing their throat</a> - avoid that.</p></li></ul><h3>Why am I hiring for a researcher?</h3><p>I want my podcast/blog to move from just asking questions about AI to actually helping answer them. But there are too many important questions, and I need a collaborator to build up context on them all, to explore dozens of fractal sub-questions, to consider the rebuttals and syntheses, and to sharpen each others thinking.</p><p>The questions I want us to explore are very broad while at the same time requiring deep technical analysis across many domains to actually answer.</p><h3>Why am I hiring this way?</h3><p>Well, I could just put out a job ad for a researcher, but I&#8217;ll get 1,000 different resumes, and I&#8217;ll have no clue based on that information whether the applicant would be any good at synthesizing lots of technical arguments and information. So I thought, let&#8217;s just list out some questions where I genuinely don&#8217;t know the answer and would be keen to get some insight.</p><h3>What this role looks like</h3><ul><li><p>Ideally in person in San Francisco, but potentially open to remote.</p></li><li><p>Will pay competitively</p></li></ul><h3>Submit <a href="https://airtable.com/app8aYOTzMkv9qeAJ/pagHhju8B5tgu4yXc/form">here</a></h3><p>If you have questions or comments, I&#8217;m hello@dwarkeshpatel.com.</p>]]></content:encoded></item><item><title><![CDATA[What I learned this week - Pretraining parallelisms, Can distillation be stopped, Mythos and the cybersecurity equilibrium, Pipeline RL, On why pretraining runs fails]]></title><description><![CDATA[April 15, 2025]]></description><link>https://www.dwarkesh.com/p/what-i-learned-april-15</link><guid isPermaLink="false">https://www.dwarkesh.com/p/what-i-learned-april-15</guid><dc:creator><![CDATA[Dwarkesh Patel]]></dc:creator><pubDate>Wed, 15 Apr 2026 14:03:00 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!QpJ5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e801a3e-5563-40fc-ba70-4af569d80647_555x772.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>At the end of <a href="https://www.dwarkesh.com/p/michael-nielsen">my conversation with Michael Nielsen</a>, we talked about how to actually retain what you learn. Michael&#8217;s advice was to make some kind of demanding artifact. Write something up. Try to explain it. So in that spirit, here are notes on some topics I&#8217;ve learned about over the last week or two. These notes are extremely rough, and have many mistakes.</p><h3>Can distillation be stopped?</h3><p>Can the frontier labs stop distillation? Because if they can&#8217;t, open source commoditizing models can catch up incredibly rapidly, making the long run business model for the labs less viable. Let&#8217;s say it takes 1T tokens from a frontier model to capture its juice (I have no idea if that&#8217;s correct, but let&#8217;s say). Even ignoring savings from caching, Opus 4.6 is $25/MTok. So $25 million for those 1T tokens. That&#8217;s nothing.</p><p>Labs are responding by hiding chain of thought. But there&#8217;s two problems with this solution:</p><ul><li><p>Chain of thought is not made of some fundamentally different kind of token. You can just instruct the model to not think first but just start solving the problem, or to write out its thinking somewhere else.</p></li><li><p>Even if labs do figure out how to robustly hide chain of thought to train in the future, you can make reconstructing the chain of thought necessary to reproduce a decoded sequence as an RLVR target. Yes that costs more, but seems doable.</p></li><li><p>Maybe most importantly, the real juice of these agentic models is their tool use (writing and updating files of code, running bash commands, etc). And if these things are done locally on the user&#8217;s computer, you can&#8217;t really hide them. And it seems like a hard lift to get users to migrate all their development workflows to a cloud that you fully control and hide visibility to, modulo a Claude agent input text prompt.</p></li></ul><p>By the way, I learned about an interesting way companies which build products atop API access to AI models can basically distill these models, in a way that potentially makes the distilled models even better than the ones they&#8217;re actually built atop.</p><p>Suppose you&#8217;ve got a coding product. In order to build a feature, a user uses your product to query some frontier model API across 10+ back and forths. Once the user is satisfied with the end result, you have the end state that the user actually wanted - &#8220;the gold diff&#8221;. These coding product companies can now set the gold diff as the RL target for training their own models, where the model gets rewarded for producing outputs that look like what users eventually converged on, and penalized for producing the kinds of intermediate outputs that users kept rejecting or editing.</p><h3>On why pretraining runs fails</h3><p>Had an interesting chat with someone on why pretraining runs often fail. It was very interesting to get a sense of all the tangible ways that things can get fucked, and why training is such a precarious operation. At a high level, breaking causality, and adding bias, seem to be key culprits.</p><p>Breaking causality:</p><ul><li><p>When you do expert routing, you first go through the router, which gives you a score of how much each token wants each expert. There&#8217;s two ways to proceed from here: 1. Token routing, where you read the scores from the token&#8217;s perspective, and allocate to each token&#8217;s top k experts. Problem is that you could end up with wildly unbalanced allocation across experts, which is terrible for performance. Alternatively, you could (and only in training) do expert choice, where you just split the tokens by which are more relatively preferred by each expert. This way you can enforce that each expert gets roughly the same number of tokens. But the big problem is that this breaks causality, because which expert token n gets allocated to may depend on which expert token n + k might be router to. And breaking causality is very bad, because you&#8217;re getting information in training (and updating based on it) that you wouldn&#8217;t see in deployment.</p><ul><li><p>Rumor is that this explains why Llama 4 was underwhelming.</p></li><li><p>I guess you could do expert choice during prefill inference? But maybe it doesn&#8217;t work well in practice to allocate tokens to experts which would not have received that token in actual training.</p></li><li><p>Tbh I don&#8217;t fully understand why breaking causality is so bad. I understand you can&#8217;t see beyond causality in real inference. But why is this minor deviation such a big issue?</p></li></ul></li><li><p>Another thing that can break causality is token dropping. Where experts just ignore the tokens in the batch that they&#8217;re supposed to process, but which rank not so strongly, and cutting whom would spare going outside padding. This breaks causality cause a later token being more strongly matched to this expert might lead to an earlier token getting ignored.</p><ul><li><p>Apparently this was an issue with Gemini 2 Pro.</p></li></ul></li></ul><p>Adding bias:</p><ul><li><p>Bias much worse than variance - variance can average out, but bias compounds</p></li><li><p>Apparently the original GPT 4 training was slow and got initially fucked because of the following bug: they were using FP16 on their collectives like all-reduce. FP16 distributes its granularity according to logarithmic density - between 1 and 2, the mantissa bits carve the interval ~0.001 apart. But 1024 and up, the mantissa might be carving the interval by multiple whole number values. Suppose some collective involves adding 1 + 1 &#8230; 10,000 times - you could get in a situation where as soon as you get to 1024, you add 1, it goes to 1025, you round down to the nearest interval at 1024, add one again. And so the calculated value is 10x off the real value. Huge issue if you&#8217;re trying to sum many small gradients into a large accumulator. And imagine how hard the bug must have been to find!</p></li></ul><p>Implications for AI training:</p><ul><li><p>Some of the people who think we can cure aging argue that there&#8217;s basically 5 different ways people die of old age (heart disease, cancer, etc), and that if we cure these 5 different diseases, then we&#8217;d basically have solved again. You could ask a similar question about these failed pretraining runs - are there 5 different ways training runs fail, in which case once a lab figures out numerics and , you&#8217;ll just have smooth sailing, or will you keep seeing new bespoke issues emerge at each new level of scale? The person I talked to seemed to think the later - he pointed out that even within numerics, there&#8217;s so many ways you can fuck things up. And new ones will keep emerging at scale.</p></li><li><p>Bearish on AI fully automating kernel writing anytime soon. Presumably this is because he thinks it&#8217;s more of an AGI complete problem than some give it credit for. There&#8217;s another school of thought that says, &#8220;Hey, which kernel gets attention or MLP to run fastest on this scaleup is a super verifiable domain, thus we can RL to superhuman performance easily.&#8221; But he says, it took Nvidia, which has the best kernel engineers in the world, a long time to optimize for Blackwell, which suggests that actually it&#8217;s quite hard, and might not be super easy to close the loop on.</p></li><li><p>Sometimes people say inference for RL generation and inference for end user generation is basically the same. But this person pointed out that in RL inference, numerical drift between inference and training engine can cause these subtle off policy biases, which matter a ton for highest quality training. But are not an issue if just serving to users.</p></li><li><p>Emphasized how important it is to have a disciplined process for amalgamating compute multipliers, because of the risks of stacking up bugs with subtle biases.</p></li></ul><h3>Pretraining parallelisms</h3><p>Notes from an excellent lecture that <a href="https://horace.io/">Horace He</a> gave my friends and me. </p><p>What made this lecture so good is that Horace built up the whole topic as a chain of problems and solutions: here&#8217;s what we want to do, here&#8217;s why it breaks, here&#8217;s how we fix it, here&#8217;s why that fix eventually breaks too. Most explanations just list out a hodge podge of different strategies, without ever connecting them to the problems they solve or explaining why you&#8217;d pick one over another.</p><ul><li><p>Equation for pretraining flops = 6ND. 2 FLOPs per parameter per token for the forward pass (multiply + add). Backward pass is 2&#215; forward because you compute gradients w.r.t. both input matrices. So 2 + 4 = 6.</p></li><li><p>Okay we can&#8217;t do all this on one GPU. So how do we split up this problem? The obvious solution is to do data parallel - where you copy the model weights across each GPU, and you just do a part of the batch on each GPU.</p><ul><li><p>The obvious problem is that each GPU only has a limited amount of HBM - B300 is 288GB - and this is not enough to store the weights as models get bigger and bigger, much less their activations.</p></li></ul></li><li><p>Okay so next thing we try is fully sharded data parallel - each GPU only stores 1/N of the parameters of each layer - before processing each layer, you all-gather the full layer&#8217;s parameters from all GPUs (each GPU only stores 1/N of each layer). After processing, each GPU discards the gathered parameters.</p><ul><li><p>It was emphasized that this is the go to default. And you only move on from this when having too many GPUs forces you to move on, for reasons explained later. The reason this is the default is that it&#8217;s trivial to overlap compute and communication time - that&#8217;s because the only thing being communicated is the weights, which are not dependent on what happened in the layer before, so you can start all gathering the next layer while you&#8217;re still computing this layer. Compare this against tensor or expert parallelism, which do need to share activations for one layer before you can process the next one. The problem with pipeline parallelism is bubbles as explained below.</p></li><li><p>From a comms volume perspective, FSDP looks insanely expensive at first &#8212; you all-gather every layer&#8217;s full weights across all GPUs, use them for one matmul, then throw them away. But this ignores what regular data parallelism already costs you - in regular DP, you still need to do an all reduce after every layer of the backwards pass in order to sync the batch&#8217;s gradients across all the GPUs. That all-reduce has comms volume of params &#215; 2. FSDP adds all-gathers &#8212; one per layer in the forward pass, one per layer in the backward pass. But an all-gather is half the comms volume of an all-reduce. So naive FSDP comms volume ends up being # params * 4 (all gather forward and back, plus all reduce on back). You can do even better: since each gradient shard only needs to end up on the one GPU that owns it, replace the all-reduce with a reduce-scatter (which skips the final broadcast step). That gets you to params &#215; 3 total &#8212; a 50% overhead over vanilla DP.</p></li></ul></li><li><p>So why can&#8217;t you always just do FSDP?</p><ul><li><p>Comms crossover: You want your compute time to be greater than your comms time - you don&#8217;t want to be bottlenecked on comms. But since compute time for FSDP decreases as you increase the number of GPUs, and comms time does not, as you scale the number of GPUs on FSDP, your MFU can totally crater. When this happens, you need to add pipeline parallelism too.</p><ul><li><p>Compute time = (6 * # tokens * active params) / (compute per GPU * number of GPUs)</p><ul><li><p>This decreases as you increase number of GPUs</p></li></ul></li><li><p>Comms time = (# total params * 3) / (nv link domain size * infiniband BW)</p><ul><li><p>Comms time does not increase as you add more domains. This was really confusing to me. Each domain collectively holds all the parameters, and you need to sync gradients across domains after each layer of the backward pass. You&#8217;d think that adding more domains means more hops in the ring, so the all-reduce gets slower. But the standard ring algorithm splits the message into one chunk per participant. More domains means more hops, but proportionally smaller chunks per hop. (This breaks down when chunks get so small that per-hop latency dominates, at which point you switch to tree algorithms.)</p><ul><li><p>Technically, you can do better than a naive single all reduce for the gradients between all the domains. You do a hierarchical collective to optimize comms time across multiple NVLink domains. Key thing to remember is that each GPU in the domain gets its own bandwidth access to infiniband. So you wanna use it all up since interconnect bandwidth is the bottleneck. You do this by trying to do as much as possible within a scaleup before you move out. So you do reduce scatter within a scale up to give each GPU the domain-level reduced gradients for a shard of the layer, then all reduce these shards across corresponding GPUs across domains, then all gather within a domain. This shifts the comms time line down, thus moving the crossover point to the right.</p></li><li><p>Made an animation to illustrate it using Cursor and Composer 2:</p><div class="native-video-embed" data-component-name="VideoPlaceholder" data-attrs="{&quot;mediaUploadId&quot;:&quot;af55c305-6da7-47cf-b799-2970cb9467e0&quot;,&quot;duration&quot;:null}"></div></li></ul></li></ul></li></ul></li></ul><ul><li><p>If you look at the equations, you can see that if you increase batch size, crossover point moves to right, and if you make the model more sparse, moves to the left.</p></li><li><p>Also why TPUs are better at FSDP - because more accelerators within a domain.</p></li></ul></li></ul><ul><li><p>Batch size floor: FSDP is data-parallel, so each GPU processes at least one sequence. Attention is computed within a sequence and can&#8217;t (easily) be split across GPUs. If your critical batch size is 10M tokens and sequence length is 10K, you only have 1K sequences &#8212; so you can&#8217;t scale beyond 1K GPUs with pure FSDP, even if you have plenty of comms bandwidth left.</p></li><li><p>Problems with pipeline parallelism (the next addition you&#8217;d make to FSDP in order to deal with these issues):</p><ul><li><p>The problem with pipeline parallelism is different - there you have bubbles that emerge from the fact that at the beginning of the batch, the GPUs dedicated to the final layers are not being used, and conversely at the end of the batch, the GPUs dedicated to the first layers are not being used. The reason you can&#8217;t overlap batches in training to solve pipeline bubbles is that you need to consolidate gradients and update the model before you process the next batch.</p></li><li><p>But also you&#8217;re adding architecture constraints - things like Kimi&#8217;s attention-to-residuals (where each block attends to all previous layers&#8217; residuals) become very difficult when those residuals live on different pipeline stages. Similarly, interleaving sliding-window and global attention layers could cause load imbalance across stages. Dealing with all this slows down research iteration, which is the greatest sin you can commit.</p></li></ul></li></ul><h3>Mythos and the cybersecurity equilibrium</h3><p>It seems like the key difference between Mythos and previous versions is that while previous versions could find individual vulnerabilities in the code (&#8220;Hey, there&#8217;s a missing bounds check here&#8221;), Mythos is long run agentic enough to rope 5 different vulnerabilities together which are all required in order to find an exploit (&#8220;Now I can execute arbitrary code, escalate privileges, etc&#8221;). To the extent that some discontinuity has been hit, it&#8217;s probably more the result of the combinatorial nature of cyberattacks rather than some off-trend increase in intelligence.</p><p>What does this mean for offense/defense? One way to look at it is that software is more secure today than it was 20 years ago, despite more and more human intelligence probing at public code, both white hat and black hat. If we get another influx of intelligence suddenly, why should the dynamic change?</p><p>In fact, we know that our foreign adversaries almost certainly have access to a bunch of critical zero days which they&#8217;re saving for a rainy day, or already using in inconspicuous ways. To the extent that Glasswing allows the whole industry to find a bunch of these latent exploits and patch them, shouldn&#8217;t we expect defense to have become much stronger relative to offense by the end of 26? Of course, this is thanks to the fact than American companies got there first and are cooperating with other companies and our government to patch things before our adversaries get to the same level.</p><p>One counterpoint I heard from a security expert is that there&#8217;s big difference between finding vulnerabilities and patching them - and AI is much better at the first than the later (people often talk about the offense/defense balance, but difficulty of finding versus patching vulnerabilities seems much more significant). In order to patch an issue, you have to find a fix that will not interfere with all the ways people use your software, and all the features which rely on weird bespoke behavior. XKCD has a nice comic illustrating how these kinds of issues come up:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!QpJ5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e801a3e-5563-40fc-ba70-4af569d80647_555x772.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!QpJ5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e801a3e-5563-40fc-ba70-4af569d80647_555x772.png 424w, https://substackcdn.com/image/fetch/$s_!QpJ5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e801a3e-5563-40fc-ba70-4af569d80647_555x772.png 848w, https://substackcdn.com/image/fetch/$s_!QpJ5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e801a3e-5563-40fc-ba70-4af569d80647_555x772.png 1272w, https://substackcdn.com/image/fetch/$s_!QpJ5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e801a3e-5563-40fc-ba70-4af569d80647_555x772.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!QpJ5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e801a3e-5563-40fc-ba70-4af569d80647_555x772.png" width="555" height="772" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5e801a3e-5563-40fc-ba70-4af569d80647_555x772.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:772,&quot;width&quot;:555,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!QpJ5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e801a3e-5563-40fc-ba70-4af569d80647_555x772.png 424w, https://substackcdn.com/image/fetch/$s_!QpJ5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e801a3e-5563-40fc-ba70-4af569d80647_555x772.png 848w, https://substackcdn.com/image/fetch/$s_!QpJ5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e801a3e-5563-40fc-ba70-4af569d80647_555x772.png 1272w, https://substackcdn.com/image/fetch/$s_!QpJ5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e801a3e-5563-40fc-ba70-4af569d80647_555x772.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Potential solutions, if it&#8217;s non-trivial to just push patches to every piece of software?</p><ul><li><p>TODO - I know nothing about formal verification of software - check out what a seL4 proof of some behavior might look like</p></li><li><p>Use LLMs to rapidly port all C to Rust. Curious how easily Mythos can find vulnerabilities in memory safe languages.</p></li></ul><p>In some sense, its good that Anthropic didn&#8217;t release this model publicly until critical IT could be patched up. In another sense, isn&#8217;t it a super bad precedent for private companies to be hoarding the ability to be able to break into any operating system and browser and device? One obvious question for Anthropic is why they didn&#8217;t just build some kind of classifier which would detect whether you&#8217;re using the model for cyberattack type stuff, and refuse requests if yes, and release that publicly.</p><ul><li><p>Patching your own software is isomorphic to finding bugs in someone else&#8217;s repo from the perspective of an LLM (and patching your own software is a frequent coding model use case).</p></li><li><p>These kinds of classifiers can be easy to evade if you have enough expertise to break the problem of finding exploits down into smaller subproblems of finding vulnerabilities which each individually seem like sensibly good behavior to an LLM with no memory</p></li></ul><h3><a href="https://arxiv.org/pdf/2509.19128">Pipeline RL</a> paper summary</h3><p>As you keep RLing a model, not only does the average length of a response increase (since you&#8217;re basically training the model to think for longer before answering) but the variance in length also increases - sometimes you get an easy problem and you can immediately answer it - other times, you need to go think for 100k tokens.</p><p>This is a big problem for GPU utilization on training. Because you have to wait for all these stragglers to finish generating before you can start the next training step.</p><p>Okay one way you could get out of this conundrum is to just to just batch generation so that while stragglers keep going, you generate even more rollouts.</p><p>The problem is that there is an optimal batch size for each training step, so you&#8217;d need to split all these rollouts you made across lots of consecutive training steps.</p><p>But this takes you into the domain of offline RL, because your model is changing with each training step. And so you&#8217;re training your model on trajectories that were actually generated by an earlier model, which is not ideal.</p><p>Pipeline RL paper proposes the following fix: in flight weight weight updates - where you just sub out the generating model partway though these generating trajectories as soon as the new training step is done, so all the short trajectories, and a good chunk of the long trajectories, that the next training step will be trained on are generated by the most recent version of the model.</p>]]></content:encoded></item><item><title><![CDATA[Notes on Space GPUs]]></title><description><![CDATA[Turning my Elon prep into a blog post]]></description><link>https://www.dwarkesh.com/p/notes-on-space-gpus</link><guid isPermaLink="false">https://www.dwarkesh.com/p/notes-on-space-gpus</guid><dc:creator><![CDATA[Dwarkesh Patel]]></dc:creator><pubDate>Thu, 05 Feb 2026 18:26:47 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/2dba4f31-2d4f-485a-835f-5c9bc75f9ce4_300x168.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p></p><p>John Collison and I <a href="https://www.dwarkesh.com/p/elon-musk">just interviewed Elon</a>. The interview was recorded before we knew that SpaceX was acquiring xAI, so the fact that our first topic was space GPUs now feels all the more relevant.</p><p>As I was preparing to interview Elon, I put together some notes and a <a href="https://docs.google.com/spreadsheets/d/1fa48HAwXaboEXNOrAj-xJF2Vv_xxQZjAtgoTu0FnnlY/edit?usp=sharing">spreadsheet</a> to help me think through orbital datacenters. I turned those notes into this blog post.</p><p>Even if orbital data centers don&#8217;t make sense yet, in the long run the singularity is clearly moving into space. Earth intercepts about one two-billionth of the sun&#8217;s total output. If AI scaling continues, compute will eventually move to where the energy is. So space GPUs are fun to think about, because they give you a sneak peek at the future. Whether that future arrives in 2030, 2040, or 2050 is another question.</p><p><strong>Please take everything below with grains of salt&#8212;grains so big that you might confuse them for rocks. Assume all the numbers are wrong.</strong> Every paragraph below covers a topic that would take an actual expert a week to properly evaluate. What you&#8217;ll find here is what a professional podcaster has pieced together from conversations with LLMs and some very generous people who talked to me before the interview. Thanks to <a href="https://x.com/CJHandmer">Casey Handmer</a>, <a href="https://x.com/PhilipJohnston">Philip Johnston</a>, <a href="https://x.com/ezrafeilden">Ezra Feilden</a>, <a href="https://x.com/andrewmccalip">Andrew McCalip</a>, <a href="https://x.com/vinayramasesh">Vinay Ramasesh</a> and the team at <a href="https://www.kineticpartners.com/">Kinetic Partnership</a> for all their help.</p><h2><strong>Why orbital data centers?</strong></h2><p>The whole reason to go to space is energy. Yes, panels in space get about 40% more irradiance&#8212;but the real advantage is that you can put your satellites in <a href="https://en.wikipedia.org/wiki/Sun-synchronous_orbit">sun-synchronous orbit</a>, where they face the sun continuously. No nights, no clouds, no need for batteries (which is the majority of cost in a solar-storage system). Solar on Earth has a roughly 25% capacity factor, meaning panels only generate a quarter of their peak output on average. In space, you get close to 100%.</p><div class="native-video-embed" data-component-name="VideoPlaceholder" data-attrs="{&quot;mediaUploadId&quot;:&quot;784c41e4-422c-4b40-848b-efaea0c75392&quot;,&quot;duration&quot;:null}"></div><p>The logic is that if the launch costs continue to drop, it will become cheaper to put GPUs in orbit than to build power plants and batteries on Earth. And there&#8217;s a lot of room for launch costs to fall&#8212;propellant is cheap, and the main expense is the rocket, which you can now reuse. Falcon 9 is around $2,500/kg with a disposable upper stage. Starship with full reusability could get below $100/kg.</p><p>But here&#8217;s the problem with this argument. Energy is only about 15% of a datacenter&#8217;s total cost of ownership. The chips themselves are around 70%. And you still have to launch those to space!</p><p>It gets worse. On Earth, GPUs fail constantly. In the <a href="https://arxiv.org/abs/2407.21783">Llama 3 paper</a>, Meta reported a failure roughly once every three hours across a 16,000 H100 cluster. When a chip dies, a technician walks over, swaps it out, and the cluster keeps running. In space, you can&#8217;t do that&#8212;at least not until we have Optimus robots stationed on every satellite.</p><p>What about radiation? It&#8217;s actually less catastrophic than you might expect. Google&#8217;s Suncatcher paper <a href="https://arxiv.org/abs/2511.19468">found</a> that their TPUs survived nearly 3x the total ionizing dose needed for a 5-year mission before showing permanent degradation.</p><p>I asked Elon about this. He responded:</p><blockquote><p>&gt; &#8220;Actually, it depends on how recent the GPUs are that have arrived. At this point, we find our GPUs to be quite reliable. There&#8217;s infant mortality, which you can obviously iron out on the ground. So you can just run them on the ground and confirm that you don&#8217;t have infant mortality with the GPUs.&#8221;</p><p>&gt; &#8220;But once they start working, their actual reliability&#8212;and you&#8217;re past the initial debug cycle of Nvidia or whatever, or whoever&#8217;s making the chips, could be Tesla AI6 chips or something like that, or it could be TPUs or Trainiums or whatever&#8212;is actually quite reliable past a certain point. So I don&#8217;t think the servicing thing is an issue&#8221;</p></blockquote><p>Consider what&#8217;s actually being proposed here. You assemble your GPUs into racks on Earth, run them for a few hundred hours to catch the duds, disassemble everything, pack it into a satellite, launch it, and get it operational in orbit. Throughout this entire process, the most expensive part of your system&#8212;the chips&#8212;are just sitting there not doing useful work.</p><h2><strong>Is this just not possible on Earth?</strong></h2><p>Throughout the interview, Elon kept returning to one point over and over again: <em>Look, forget the economics! It will simply not be physically possible to scale power production to the scale needed for AI on Earth. </em>He went on:</p><blockquote><p>&gt; &#8220;The only place you can really scale is space.&#8221;</p><p>&gt; &#8220;All of the United States currently uses only half a terawatt on average. So if you say a terawatt, that would be twice as much electricity as the United States currently consumes. So that&#8217;s quite a lot. Can you imagine building that many data centers? That many power plants? It&#8217;s like those who have lived in software land don&#8217;t realize they&#8217;re about to have a hard lesson in hardware.&#8221;</p></blockquote><p>Elon kept pointing out the bottlenecks we&#8217;ve already run into on Earth. You can&#8217;t plug into the utilities&#8212;the interconnect queues are too long. You can&#8217;t do behind the meter and generate power yourself&#8212;lead times for turbines stretch past 2030. You can&#8217;t do solar on Earth, because of permits, and because of the tariffs. And Earth has clouds and nights, requiring overbuilt solar and batteries. In space, you can just put the satellites in sun synchronous orbit!</p><p>Look, at some level, <em>it</em> is true that we can&#8217;t keep scaling on Earth. But keep in mind that the Earth is really fucking big. 1 TW of solar (with 25% capacity factor, so really 4 TW of panels) is around 30,000 square miles. That&#8217;s like 1% of the US&#8212;about the size of South Carolina. For context, AI datacenters currently consume only ~20 GW globally.</p><p>By the time we&#8217;re talking about multiple terawatts, we&#8217;ll have had to massively scale leading-edge wafer production. And that&#8217;s the really hard part. Fabs are the most complicated manufacturing facilities humans have ever built. In order to believe that we need to go to space in order to find the power turn on all these chips, we&#8217;ll need to assume a few things:</p><ul><li><p>We&#8217;ll manage to produce <em>a lot </em>more chips.</p></li><li><p>Every single relief vessel for power generation on Earth will fail to scale.</p></li></ul><p>But semiconductors are so much more complicated than solar panels! They&#8217;re even more complicated than the blades on a turbine. It feels quite unlikely to me that the thing we manage to solve is building terawatts worth of leading edge wafers, but in that world we can&#8217;t figure out how to pave Nevada (or if regulation proves to be a problem, then the UAE) with solar panels.</p><h2><strong>100 GW into space</strong></h2><p>How many Starship launches will it take to launch a 100 GW into space?</p><p>An orbital datacenter satellite has three big components: solar arrays, computers, and radiators. And the key constraint is that for every watt of compute, we need roughly one watt of solar and one watt of thermal rejection capacity.</p><p>The W/kg of each component determines how the mass budget gets split&#8212;and how much compute you can bring along. The figure that matters most here is the specific power of the whole satellite: after you account for solar panels, radiators, and chassis, how many watts of compute do you actually get per kilogram launched?</p><p>For Starlink satellites, this works out to roughly 50 W/kg. The people trying to build orbital datacenters are currently targeting 100 W/kg. There are only two ways to get there: lighter solar panels (more watts generated per kg) or lighter radiators (more watts rejected per kg).</p><p>The numbers below are super rough. Reliable figures for space-grade components are hard to come by. But even rough math reveals which variables must improve&#8212;and by how much&#8212;in order to hit 100 W/kg.</p><ul><li><p>Solar: There are apparently companies that are targeting next gen thin film that reaches <a href="https://news.satnews.com/2026/02/01/orbital-vs-terrestrial-solar-the-math-of-energy-density-and-capacity-factors/">upwards of 500 W/kg</a>, but the state of the art is <a href="https://en.wikipedia.org/wiki/Space-based_solar_power">150 W/kg</a>, and most missions right now fly <a href="https://www.nasa.gov/smallsat-institute/sst-soa/power-subsystems/">30 W/kg</a>. Let&#8217;s be generous and assume 200 W/kg.</p><ul><li><p>The trouble here is that there&#8217;s obviously a tradeoff&#8212; denser panels costs more money, but reduces launch cost. And it&#8217;s difficult to calculate what that implies for these next gen panels, because their prices are not listed anywhere.</p></li></ul></li><li><p>Compute: I&#8217;ve heard that a stripped down GB200 NVL72 with no cooling equipment is around 100 kg. They draw 132kW of power, but let&#8217;s add 10% overhead for the intersatellite lasers and so on. That gets us to 1,452 W/kg.</p></li><li><p>Radiators: In space, you can&#8217;t convect heat away, because there&#8217;s no air. You can only radiate it, which means your panels glow infrared until the heat leaves. The Stefan-Boltzmann law governs how much power a surface can radiate.</p><p>GPUs typically run up to 90&#176; Celsius. There&#8217;s some temperature drop through the heat pipes and fluid loops that carry heat to the radiator surface. Call it 30&#176;C. So your radiators end up operating around 60&#176;C. Plug that into Stefan-Boltzmann (assuming you&#8217;re using aluminum panels that weigh around 2 kg per square meter of surface area, that works out to roughly 320 W/kg.</p><p>Since radiated power scales with T&#8308;, running your chips hotter can help you save a lot of radiator mass. For space, people will have to figure out how to build GPUs that tolerate higher temperatures.</p></li></ul><p>Assuming the numbers above&#8212;and also assuming that a fourth of the mass of the satellite has to be the chassis&#8212;I get 85 W/kg for the whole system. Again, I want to emphasize these are <em>rough</em> calculations; feel free to plug in your own numbers in the spreadsheet <a href="https://docs.google.com/spreadsheets/d/1fa48HAwXaboEXNOrAj-xJF2Vv_xxQZjAtgoTu0FnnlY/edit?usp=sharing">here</a>.</p><p>At 150 metric tons to low earth orbit per Starship (Elon&#8217;s target), you&#8217;re looking at around 10 MW per launch. That means roughly 100 Starship launches in order to put 1 GW of compute in orbit. To hit 100 GW in a year, you&#8217;d need roughly 10,000 launches, or, about one launch every hour.</p><p>This is insane! A single Starship produces around 100 GW of thrust power at liftoff. That&#8217;s about a fifth of total US electricity consumption, concentrated in one rocket for a few minutes. And the plan would be to do that once an hour, every hour, every day, for a year.</p><p>I asked Elon what that world looks like:</p><blockquote><p>I don&#8217;t think we&#8217;ll need more than... I mean, you could probably do it with as few as 20 or 30 [Starship vehicles]. It really depends on how quickly the ship has to go around the Earth and the ground track before the ship has to come back over the launch pad. So if you can use a ship every, say, 30 hours, you could do it with 30 ships. But we&#8217;ll make more ships than that. SpaceX is gearing up to do 10,000 launches a year, and maybe even 20 or 30,000 launches a year.</p></blockquote><h2><strong>Workloads and comms</strong></h2><p>Starlink satellites already communicate via inter-satellite laser links <a href="https://en.wikipedia.org/wiki/Laser_communication_in_space#cite_note-49">at 100 Gbps</a>&#8212;and Google&#8217;s Suncatcher paper suggests off-the-shelf transceivers could potentially hit 10 Tbps. For context, Infiniband links between nodes in a terrestrial datacenter run <a href="https://marketplace.nvidia.com/en-us/enterprise/networking/400gbeosfpcables/">at 400 Gbps</a>. The gap isn&#8217;t as large as you might expect. So, could you do synchronous training in space?</p><p>Even the most bullish analysts don&#8217;t claim that orbital data centers will be used for training. I don&#8217;t know any of the relevant orbital mechanics, but obviously satellites at different altitudes move at different orbital velocities, which means the satellites are desyncing relative to one another. Google came up with a clever solution for this in their Suncatcher paper&#8212;keep lots of satellites in a single tight cluster at the same altitude. Google&#8217;s researchers proposed eighty-one satellites in such a synchronized constellation. If each constellation had a GB200 NVL72, then that&#8217;s only 15 MW parcels of coherent compute.</p><p>Defenders of orbital datacenters say that most compute is going to shift to inference (and with RL, most training is also inference). Maybe the legacy terrestrial datacenters do end up doing the pretraining runs, and then whatever mixture of RL environment training and continual learning  happens in the future does happen in space. So, the argument goes, it&#8217;s not a big deal that the scale ups in space are isolated. But there&#8217;s still the question of how hundreds of gigawatts of inference are beamed back to Earth.</p><p>For a moment, let&#8217;s imagine a world where as we see the sunrise and sunset we also see a Saturn-like belt of GPU satellites passing over us. That&#8217;s already really cool. But then there&#8217;s  another sci-fi premise, which I really wanted to be plausible, and which turns out not to make any sense: Imagine that every 12 hours, as this country of geniuses in space passes over us and shoots down half a day&#8217;s worth of new ideas, our code finally starts working and our factories buzz alight and become more productive. Unfortunately, it&#8217;s just science fiction. Inference doesn&#8217;t take that much bandwidth. One hundred gigawatts of a 5T model is roughly 58 billion tokens per second, resulting in ~ <a href="https://claude.ai/share/9b4bff2b-d114-4421-9cbf-0eff30112a3a">230 GB/s</a>.</p><p>That&#8217;s nothing. That many tokens can easily be beamed using lasers from GPUs in the orbital plane through to Starlink satellite network and then down to Earth.</p><p>Latency might be an issue, up to fifty milliseconds from any given spot on Earth through the Starlink network to the sun synchronous orbit and then back again. But as we move towards a world of true remote coworker AIs, where the agent works for tens of minutes before coming back to us, the marginal milliseconds of latency matter less and less.</p><h2><strong>So why is Elon doing this?</strong></h2><p>I&#8217;m willing to accept Elon&#8217;s argument that if launch costs become sufficiently cheap <em>and</em> we can repair GPUs in space, then there&#8217;s a viable path toward orbital data centers. But it seems especially difficult to imagine a situation in which orbital data centers end up <em>significantly </em>cheaper, because, again, most of the cost of a data center is the GPUs.</p><p>For most compute to shift to space, all of the following things would need to be true:</p><ul><li><p>Power generation on Earth hits a ceiling, or AI demand outstrips every terrestrial option.</p></li><li><p>Chip production scales faster than anyone expects, so we have the silicon but not the electricity.</p></li><li><p>Starship reaches thousands of launches per year.</p></li></ul><p>If Elon&#8217;s right, he wins the AI race outright. SpaceX is the only entity that can launch at that scale. xAI would have unlimited power. Everyone else will be stuck fighting over grid interconnects and turbine orders.</p><p>And if Elon&#8217;s future doesn&#8217;t materialize? xAI is just another lab in the pack. Which means xAI loses. The AI race is a winner-take-all competition, and xAI isn&#8217;t in first place. Elon&#8217;s comparative advantage was never going to be navigating utility interconnect queues or filing permits faster than Google. His advantage is SpaceX. So why not bet on the world where SpaceX becomes the kingmaker?</p><p>This might sound reckless. But that&#8217;s how SpaceX got here. Their whole business plan seems to be one in which they conjure new wells of demand for each generation of rocket on the path to the Dyson swarm. Falcon 9 first flew in 2010. Starlink didn&#8217;t launch until 2019. Maybe orbital datacenters end up being for Starship what Starlink was for Falcon 9.</p><p>Sometimes, during the interview, I found my thoughts drifting toward Elon&#8217;s vision for this big, interconnected future. So I paused a moment and said:</p><blockquote><p>What I find remarkable about the SpaceX business is the end goal is to get to Mars, but you keep finding ways on the way there to keep generating incremental revenue to get to the next stage and the next stage.</p></blockquote><p>Elon nodded his head slowly. And then he said:</p><blockquote><p>You can see how this might seem like a simulation to me.</p></blockquote><h2></h2><p></p>]]></content:encoded></item><item><title><![CDATA[Hiring scouts to help me find guests]]></title><description><![CDATA[$100/hour, fully remote. Ideal candidate is maybe a grad student/post doc/or working in one of: bio, history, econ, math/physics, AI/hardware.]]></description><link>https://www.dwarkesh.com/p/hiring-scouts-to-help-me-find-guests</link><guid isPermaLink="false">https://www.dwarkesh.com/p/hiring-scouts-to-help-me-find-guests</guid><dc:creator><![CDATA[Dwarkesh Patel]]></dc:creator><pubDate>Thu, 15 Jan 2026 16:02:50 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/bbfbfbf2-988e-4a86-9603-27a999afdc10_360x360.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>My main bottleneck is finding excellent guests. So, I&#8217;m hiring a couple part time scouts to help me find the next David Reich/Sarah Paine/Adam Brown.</p><p>$100/hour, fully remote, work hours are flexible - I expect it&#8217;ll be 5-10 hours a week.</p><p>Ideal candidate is maybe a grad student, or a post doc, or working in one of the fields I wanna find guests in. I&#8217;m looking for people who are really plugged into some discipline and have high taste.</p><p>Beyond just scouting guests, I&#8217;ll want your help assembling curriculums that help me prep for interviews and rapidly get up to speed.</p><p>The application form is <a href="https://airtable.com/appXzMS36pX3XAYV6/pagT9mTdjxxslroks/form">here</a>, and it&#8217;s extremely simple - just pitch me on a guest and tell me a bit about yourself. Please submit by 11:59 PM Pacific, Friday, Jan 23.</p><p>I&#8217;m looking to hire ~one scout for each of the following fields: bio, history, econ, math/physics, AI/hardware.</p><p>However, it&#8217;s very possible I end up hiring more (or fewer), or break apart the domains of knowledge in a different way, based on the range of expertise of the best people who apply.</p><h3><strong>What I&#8217;m looking for in guests</strong></h3><p>I&#8217;m looking for people who are deep experts in at least one field, and who are polymathic enough to think through all kinds of tangential questions in a really interesting way. </p><p>So I&#8217;m selecting for this synthetic ability to connect one&#8217;s expertise to all kinds of important questions about the world - an ability which is often deliberately masked in public academic work. Which means that it can only really come out in conversation.</p><p>That&#8217;s why I want to hire scouts. I need their network and context - they know who the polymathic geniuses are, who gave a fascinating lecture at the last big conference they attended, who can just connect all kinds of interesting ideas in the field together over conversation, etc.</p><p>We get tons of inbound from people who are working on impressive companies or doing interesting research projects. But almost always it&#8217;s a no; while I think their work is important, it&#8217;s self-contained in a way that I worry won&#8217;t lead to interesting broad discussion.</p><p>To get a little more concrete, here&#8217;s what worked well about some of my recent favorite guests:</p><p>Let me talk through why I think some interviews worked especially well, so you can think about what people in fields you&#8217;re familiar with fill a similar mold.</p><ul><li><p><a href="https://youtu.be/XCLODgdCmKA?si=gzt3Kvs2N4v8DTvf">Jacob Kimmel</a>: A lot of people who pitch themselves as guests are capable of only talking about their own research. But the amazing thing about Jacob is that he is an insane polymath. For example, he could explain why evolution didn&#8217;t select for longevity by drawing deep analogies to how gradients flow in ML models. He had all these other random interesting takes, from why humans never evolved their own antibiotics to how there&#8217;s this gene that used to protect us from HIV-like viruses but got repurposed, which hints at some ghost scourge. And then he could zoom out and give a great diagnosis of what&#8217;s bottlenecking pharma progress. I really want to emphasize how that&#8217;s different from other brilliant people I get pitched &#8211; these people are also doing incredible research, but they don&#8217;t have this range of really deep, interesting takes. That part is super crucial.</p></li><li><p><a href="https://youtu.be/Uj6skZIxPuI?si=g30p79rTnhTMZ6n0">David Reich</a>: It&#8217;s actually quite surprising that my second most popular guest of all time is a geneticist of ancient DNA. How did that happen? Here&#8217;s why I think this episode blew up. In high school, you get some vague explanation of human evolution. And you feel like you understand it and can move on with your life. And here comes David, showing you how this very fundamental topic, which you assume was settled and haven&#8217;t bothered thinking about in years, is actually way more murky and surprising than you realized, and how new discoveries are totally overturning our basic understanding of the field (in this case, the how, when, where of human evolution)<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a>.</p></li><li><p><a href="https://youtu.be/lXUZvyajciY?si=s8J640JwRtFCdBaX">Andrej Karpathy</a>: It&#8217;s extremely rare to get someone who is expert-level in a technical, fast-moving, and frothy field, but who has no vested interest in a particular company or approach, and who is in a position to just give an unbiased lay of the land. I have a couple questions below about biotech or formal math or robotics in the wake of AI progress - if there&#8217;s a Karpathy-type person in those fields, I&#8217;d be very keen to get a technical lay of the land and vibe check of what claims are credible versus crazy.</p></li></ul><h3><strong>Some recent questions</strong></h3><p>In case it&#8217;s helpful for brainstorming a guest, I&#8217;ve listed out a few big questions that have been on my mind recently. But please feel free to ignore them - there&#8217;s way more interesting questions in the world than the ones I am aware of - feel free to say, &#8220;You might not yet be curious about antibody development/the history of language/the dark ages/battery tech, but the guest I have in mind for that topic is so good that it&#8217;s going to be your next big banger episode.&#8221;</p><h4><strong>Bio</strong></h4><ul><li><p>Dario&#8217;s <a href="https://www.darioamodei.com/essay/machines-of-loving-grace">Machines of Loving Grace</a> argues we&#8217;ll compress a century of bio progress into a few years - that big breakthroughs like CAR-T therapy, mRNA vaccines, cheap genome sequencing, etc  show how in the long run things which seem like data or physical bottlenecks can be solved by better tools to measure/predict/perturb/understand biological system, and these tools are downstream of intelligence. But here&#8217;s what I don&#8217;t fully understand: over the last 3 decades, we&#8217;ve seen a million-fold reduction in genome sequencing costs, 1000-fold decrease in DNA synthesis costs, the development of precise gene editing tools like CRISPR, and the ability to conduct massively parallel experiments through multiplexing techniques. But it doesn&#8217;t seem like we&#8217;re curing diseases or coming up with new treatments at a faster rate now than we were 30 years ago. If anything, drug development is <a href="https://en.wikipedia.org/wiki/Eroom%27s_law">slowing down</a>. I want to find a biology researcher who can think through how plausible a 10x or 100x speedup in new drug discovery actually is. They should obviously know a lot about and have hot takes on what&#8217;s actually bottlenecking progress today, and they should be flexible enough to imagine what might change with much more intelligence.</p></li><li><p>What exactly is the special sauce of the brain that we&#8217;re still missing? <a href="https://youtu.be/_9V_Hbe-N1A?si=wSP8VXHBbzs4hOyV">Adam Marblestone thinks</a> it&#8217;s the curriculum of reward functions and the learning/steering subsystems. Others argue that gradient descent is fundamentally worse than how the brain learns within a lifetime (which is closer to in-context learning in its flexibility and sample efficiency).</p></li></ul><h4><strong>Math/Physics</strong></h4><ul><li><p>I&#8217;ve been really enjoying Strogatz&#8217;s <a href="https://www.amazon.com/Nonlinear-Dynamics-Student-Solutions-Manual/dp/0813349109/">Nonlinear Dynamics and Chaos</a> textbook, and I want to make something podcast-shaped out of it. Strogatz himself has deferred until after he finishes his next book, so I&#8217;m looking for another mathematician on a related topic. I think the right format here isn&#8217;t a normal meandering interview - it&#8217;s something more like a lecture. A mathematician comes in with a specific topic or example we can deep dive on. He posts up at a blackboard, starts explaining a topic, and I interrupt to clarify confusions and ask follow-up questions. The model is something like Terence Tao and Grant Sanderson&#8217;s<a href="https://www.youtube.com/watch?v=FPl_rag0yAo"> cosmic distance ladder video</a>. Who can replicate something similar with me with some independently explainable topic in chaos/nonlinear dynamics or adjacent topics? I&#8217;d be especially keen if someone can present something on how the topics in this textbook tie into ML (see for example<a href="https://sohl-dickstein.github.io/2024/02/12/fractal.html"> Neural network training makes beautiful fractals</a>).</p></li><li><p>What real world impact should we expect from the current batch of AI for math projects? What are the fields of technology where people are going, &#8220;Ah we could totally solve quantum computing (or fusion or AGI) only if we had more theorems!&#8221; But maybe problems in biology and physics and materials and so on reduce down to math in a way I&#8217;m not foreseeing, and automating formal math alone is enough to unlock a bunch of progress. See footnotes for some more questions I wanna ask the right guest on this topic.</p><p>I started reading<a href="https://www.amazon.com/Proofs-Refutations-Mathematical-Discovery-Philosophy/dp/1107534054"> Proofs and Refutations</a>, which is this famous 1976 book by the Hungarian mathematician Imre Lakatos about the philosophy of mathematics. He says math involves a lot of changing definitions and swapping lemmas in order to deal with different counterexamples. This seems fine for a good faith mathematical community, but super reward hackable for these AI-for-math models. Also it involves a lot of realizing how a problem in one domain is really a problem in another, and noticing the meta level pattern - AIs so far have been especially bad at this kind of thing. If math is just proof search within a fixed formal system, then AI can help a lot. But if its dialectical construction and refinement of concepts (based on what tasteful parsimonious definition can withstand counterexamples) , then I feel self play and &#8216;automated cleverness&#8217; alone won&#8217;t do the trick. But maybe automated counterexamples are super useful. I&#8217;m sure for practicing mathematicians there&#8217;s a bunch of stuff that&#8217;s naive or wrong about the above. Would love to chat out what the actual research math process is like, and what good it would do to automate it.</p></li></ul><h4><strong>AI/hardware</strong></h4><ul><li><p>RL progress has been very fast, but it&#8217;s partly the result of going from almost nothing to 1e26 FLOPs training compute in a year (aka like going from GPT-1 to GPT-4.5). It&#8217;s still possible that it has terrible scaling exponents and further progress will be very slow. And also it&#8217;s not clear how much of the progress over the last year comes from inference scaling, which has worse variable economics. But on the other hand, maybe there&#8217;s a ton of low hanging fruit in improving RL - with pretraining, there&#8217;s been 5 years of developing the theory and empirics of optimal batch sizes, learning rates, architectures, etc. As that low hanging fruit is picked, maybe RL progress continues to be fast? The other big question about RL training is how much transfer learning are we seeing - is there all this crazy meta learning that&#8217;s not directly induced by any env and which will enable flexible human-like labor soon? I have no idea. My friends at labs who are actually doing this training obviously wouldn&#8217;t tell me. But I want to actually concretely understand what&#8217;s going on here.</p></li></ul><h4><strong>History</strong></h4><ul><li><p>There&#8217;s the famous<a href="https://en.wikipedia.org/wiki/Joseph_Needham#The_Needham_Question"> Needham question</a>, which asks why China didn&#8217;t industrialize first despite leading the world in population, inventions, and bureaucratic sophistication. I find the standard explanation of how this centralized Ming/Qing regime damped invention and exploration unsatisfying. Or at least I don&#8217;t understand it concretely. It&#8217;s such a big country - how can you retard progress across the whole thing, especially given that state capacity was presumably weaker in the past? Or at least I assume it was - what did a provincial bureaucrat actually do day-to-day? Was there a price system? Private property? How did the state actually interfere with merchants and artisans?</p></li></ul><h4><strong>Economics</strong></h4><ul><li><p>There&#8217;s something unsatisfying about the <a href="https://epoch.ai/blog/explosive-growth-from-ai-a-review-of-the-arguments">arguments</a> that we&#8217;ll see 20%+ explosive economic growth from AI. Even if true, what does that mean? What is actually happening? I thought <a href="https://www.darioamodei.com/essay/machines-of-loving-grace">Machines of Loving Grace</a> was a great account of what plausibly is happening on the human facing side of the singularity - aka the FLOPs that are going towards curing disease. But presumably most of what is happening is investment towards more robots, more compute, etc. My sense of what that side of things looks like is so murky and handwavy. There is a version of Machines of Loving Grace you can do that is somewhat concrete about all the sci fi shit - not just gesturing at the galaxies, but getting specific about the space GPUs and factorio like solar tiling and all the other things I&#8217;m not thinking of which are relevant to understanding 2040. Presumably the right guest is someone who is really strong in engineering/physics and economics and has a penchant for sci-fi and has a lot of concrete ideas here.</p></li><li><p>What should India or Nigeria or for that matter any country not directly in the semiconductor/foundation model supply chain do right now? If the main mechanism of catchup growth goes away (namely, that the underutilized labor of developing countries can rapidly be made more productive with capital and know-how from the developed world), what happens to all these countries that are not China or the US?</p><p></p></li></ul><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Just to give you a sample of some of the surprising findings that he talked through:</p><ul><li><p>70,000 years ago, half a dozen different species of humans (Neanderthals, Denisovans, &#8216;Hobbits&#8217;, etc) lived across Eurasia. And then some small group of modern humans (only 1,000 to 10,000 people) drove all of them to extinction. Everyone native to Eurasia and America is descended from this one tribe.</p></li><li><p>Neanderthals may have gotten 30-70% of their DNA from modern humans. Which implies that maybe non-Africans today are actually &#8220;Neanderthals who became modernized by waves and waves of admixture&#8221; rather than modern humans with a bit of Neanderthal mixed in.</p></li><li><p>Yersinia pestis (bubonic plague bacteria) may have killed a quarter to half of all people in Western Eurasia for thousands of years, starting around 5,000 years ago. And may be central to explaining everything from the Yamnaya expansion to the fall of Rome to the Industrial Revolution.</p></li><li><p>It&#8217;s not clear modern humans were even primarily in Africa during the key period (2 million to 500,000 years ago) when human brains diverged from those of other species. Our lineage may have resided in Eurasia for significant stretches.</p></li></ul><p>Okay I&#8217;ll stop, but you see my point. What are the other fields like human evolution, and the other presenters like David Reich, who will make you go, &#8220;What the fuck, I had no idea.&#8221;</p><p>David being David is actually a huge piece of the puzzle here which I want to replicate. He&#8217;s just incredibly deep and polymathic on what may from the outside look like one field but is in fact very many, from population genetics to archeology to linguistics. And while he&#8217;s intellectually humble enough to make qualifiers, he will (and this is very important) go ahead and give hot takes and start speculating about connections between fields and how different hypotheses relate to each other and so on. He won&#8217;t just stay at, &#8220;Our results show a genetic cline between North and South Indians.&#8221; He&#8217;ll say, &#8220;And we could be wrong here, but this suggests that the caste system which enforced this never otherwise seen levels of endogamy has been incredibly strong for millennia.&#8221;</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[What I've been reading recently - Jan 10, 2026]]></title><description><![CDATA[Nonlinear dynamics and Chaos, Machines of Loving Grace, Max Hodak&#8217;s theory of consciousness, Neural network training makes beautiful fractals]]></description><link>https://www.dwarkesh.com/p/notes-jan-10-2026</link><guid isPermaLink="false">https://www.dwarkesh.com/p/notes-jan-10-2026</guid><dc:creator><![CDATA[Dwarkesh Patel]]></dc:creator><pubDate>Sat, 10 Jan 2026 20:30:07 GMT</pubDate><enclosure url="https://substackcdn.com/image/vimeo/w_728,c_limit,d_video_placeholder.png/903855670" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I was recently chatting with a friend who has a similar job to mine. We were talking about how even though our jobs are fundamentally about learning about stuff, our time so easily gets sucked up by other things. So to hold myself accountable, I&#8217;m gonna try to publish a blog post every two weeks or so where I explain what I&#8217;ve been reading.</p><h2>Max Hodak&#8217;s theory of consciousness</h2><p>I&#8217;m totally gonna butcher this - please excuse. If you wanna get the real deal, go check out his <a href="https://maxhodak.com/nonfiction/2025/12/05/the-binding-problem">summary blog post</a> and his <a href="https://youtu.be/DI6Hu-DhQwE?si=gY5YfwvqwgNGOHNk">full talk</a> on this topic.</p><p>Max is focused on two big sub-questions which together form &#8220;the binding problem&#8221;:</p><ul><li><p>Mode binding: how do color, shape, texture, and motion get combined into a unified visual percept of &#8220;a red cup&#8221;?</p></li><li><p>Moment binding: why do we experience all the neurons firing across our entire brain over the course of 10s of milliseconds as a single quanta of experience?</p></li></ul><p>Max thinks each of these binding sub-problems is related to a brain wave:</p><ul><li><p>Gamma waves - 40 Hz - Fast, local coordination of nearby neurons to get on the same page about what they&#8217;re representing.</p></li><li><p>Alpha waves - 10 Hz - Slower waves that run through the whole brain and unify experience - think of these like the forward pass of the brain.</p><ul><li><p>Two cool things about alpha waves I hadn&#8217;t realized. 1. that neurons ride the peak of this oscillation 2. when alpha waves slow down or speed up (fight or flight reactions, etc), people experience time dilation.</p></li></ul></li></ul><p>Anyways, Max points out that the brain is storing a bunch of structured representations about the world physically, and some feedback controller has to go in and make sure that these representations are correct. This is part of what the alpha waves are doing. And this feedback control and binding is consciousness. I&#8217;m glossing over a bunch of logical connections that I definitely don&#8217;t understand. But I&#8217;ll leave it here.</p><p>I know Max could provide a really good answer, but just talking to myself, I&#8217;m confused on what the reason is to think that feedback control = consciousness? By this logic, does <a href="https://en.wikipedia.org/wiki/Memory_refresh">memory refresh</a> = consciousness too?</p><p>Max thinks that figuring out what&#8217;s up with consciousness will mean discovering new physics. And specifically, physics at the level of the 4 fundamental forces - some property as basic as mass or charge. His logic is that either consciousness has no real impact on the world (it&#8217;s just a byproduct of other stuff the brain does), which would be odd, or it actually has an effect, which would mean it&#8217;s new physics.</p><p>I&#8217;m not sure I buy this. 1. Can&#8217;t it be an effect that&#8217;s best understood at an implication of existing laws of physics - the fact that wood floats on water has an impact on the world, but you don&#8217;t need new physics to explain it 2. Doesn&#8217;t it seem implausible that evolution blindly stumbled upon and is now making good use of a whole undiscovered physical field which we have never managed to actually interact with using our technology, nor seen summoned anywhere else in the universe?</p><h2><a href="https://www.amazon.com/gp/product/0367026503">Nonlinear dynamics and Chaos</a> by Steven Strogatz</h2><p>I&#8217;m only 3 chapters in, so I&#8217;ve only got the building blocks so far. The fundamental idea is this. It&#8217;s often hard to anticipate how a system will evolve just by observing a bunch of different trajectories over time. But it&#8217;s much easier to see what will happen if you plot how the system will evolve from different starting points. The examples get more and more interesting, and because Strogatz focuses on the graphical and geometric interpretations, the motivating problems are super satisfying; the book is really a bunch of 3Blue1Brown videos on a certain topic stapled together.</p><p>Side note: I could not have understood anything here if I didn&#8217;t have LLMs and couldn&#8217;t watch the lectures async. I paused every minute or so (to clarify some confusion with a chatbot or to try and anticipate the next step), and I had the same section of textbook open at the same time.</p><p>I&#8217;m now wondering to myself, &#8220;How the hell did I learn anything in college at all?&#8221; I would be so lost if I was actually taking this course in college and just attending the lectures live.</p><p>In college, I actually did bounce out of a difficult course I feel like I could totally learn today with LLMs and async lectures + my adult executive function.</p><div><hr></div><p>As I was working through these examples (some inspired by actual papers), I kept thinking about what parts the &#8220;automated cleverness&#8221; (Terry Tao&#8217;s term) of today&#8217;s AIs could actually help with.</p><p>It&#8217;s crazy how much understanding you can get about a physical system through mathematics. But that understanding is so dependent on insight and interpretation.</p><p>To give one example, Section 3.7 has a really clever model of an insect outbreak, showing how budworms, birds, and trees play out against each other given different growth rates and other dynamics.</p><p>But first you have to figure out the right dimensionless forms. And that requires judgment about which dimensions actually matter. In the insect model, the choice was to think in terms of R and K and treat the bird population as basically an artifact of those parameters. But you could have done it the other way around&#8212;from the basis of birds.</p><p>Then there&#8217;s how you make the visualization. Once you&#8217;ve got the dynamics in dimensionless form, you could just graph the equation and find the fixed points. But the result would be almost impossible to interpret. Graph it a different way, though, and suddenly the intercepts align with your intuition. You can actually <em>see</em> the three regimes: where carrying capacity is so low the population never gets going, where birds keep things in check, and where the outbreak has outgrown the birds&#8217; ability to control it.</p><p>This kind of insight is inseparable from understanding what you&#8217;re even trying to learn about the system. And I&#8217;m skeptical today&#8217;s AI helps much here. When these methods were first developed, the right forms and interpretations weren&#8217;t obvious. The mathematician who wrote the original paper had to come up with new insights about <em>how to think</em> about the problem.</p><p>Maybe models are now good enough to apply these methods to new systems that fit the same template. But that just means the few mathematicians who invent genuinely new frameworks are the only ones who stay relevant.</p><h2><a href="https://www.darioamodei.com/essay/machines-of-loving-grace">Machines of Loving Grace</a> by Dario Amodei</h2><p>Starting with the biology section: Dario argues that we&#8217;ll get a century of bio progress in a few years. His argument:</p><ul><li><p>Most bio progress is driven by breakthrough discoveries which give you whole new primitives for what you can measure, change, or predict (CAR-T therapy, mRNA vaccines, CRISPR, genome sequencing costs declining so much, etc).</p></li><li><p>These discoveries seem to have been made in scrappy haphazard ways, often years after they were initially possible, and often by people responsible for other breakthroughs as well. All 3 of these observations hint that they are bottlenecked by intelligence.</p></li><li><p>Dario acknowledges that data is a huge bottleneck for bio. But the tools we have for collecting data can also be expanded by intelligence. Human researchers came up with multiplexing and AlphaFold and Perturb-Seq - the AI researchers will come up with even more.</p></li></ul><p>Here&#8217;s the counterargument. The kinds of human researcher breakthroughs he uses as examples of what AI could do more of haven&#8217;t had a huge impact on health. Over the last 3 decades, we&#8217;ve seen a million-fold reduction in genome sequencing costs, 1000-fold decrease in DNA synthesis costs, the development of precise gene editing tools like CRISPR, and the ability to conduct massively parallel experiments through multiplexing techniques. But it doesn&#8217;t seem like we&#8217;re curing diseases or coming up with new treatments at a faster rate now than we were 30 years ago. If anything, drug development is slowing down. Why think that AI will be able to fundamentally change this dynamic?</p><p>Relatedly, Jacob Trefethan has an excellent <a href="https://blog.jacobtrefethen.com/ai-san-francisco/">blog post</a> makes the the argument that AI won&#8217;t speed up medical progress that much (he also steelmans the opposite point in <a href="https://blog.jacobtrefethen.com/ai-optimism/">this other post</a>). Jacob points out that making a drug to cure something like Alzheimer&#8217;s is really hard. Raw understanding of some of the disease life cycle (which more intelligence could give you more of) is not enough. We understand that Alzheimer&#8217;s is clearly linked to Amyloid beta, and there are now many different drugs trying to remove amyloid plaques which have all not worked. Even if we get more insights like the Amyloid beta thing from AI scientists, that alone will not be enough to identify the correct targets. You just have to do a bunch of experiments on live humans.</p><p>This is why Dario&#8217;s point about clinical trials falls flat. He argues that clinical trials are currently slow because we just don&#8217;t know whether a given drug will actually work. But if we had much greater confidence, like we did with the mRNA vaccines for COVID, then we could test and approve drugs much faster. However, I don&#8217;t see why we should think that modulo the full hyperrealistic simulation of the human body, we <em>could</em> tell ex ante which drugs are gonna work. I don&#8217;t yet buy the argument that a million George Church clones in a datacenter could derisk all the drug trials</p><p>Quick notes on other parts of the essay:</p><ul><li><p>Overall I find it pretty impressive that a tech CEO is this generally thoughtful.</p></li><li><p>The poverty and econ section doesn&#8217;t address that the main mechanism of catchup growth goes away post AGI; namely developing countries have lots of underutilized labor which is bottlenecking production, and because the marginal product of labor is high in the world today, those countries can get rich fast. So how exactly are these other countries catching up?</p></li><li><p>The key point that underlies his framework that intelligence can drive a century of progress in 5-10 years : &#8220;Things that are hard constraints in the short run may become more malleable to intelligence in the long run. For example, intelligence might be used to develop a new experimental paradigm that allows us to learn <em>in vitro</em> what used to require live animal experiments, or to build the tools needed to collect new data (e.g. the bigger particle accelerator), or to (within ethical limits) find ways around human-based constraints (e.g. helping to improve the clinical trial system, helping to create new jurisdictions where clinical trials have less bureaucracy, or improving the science itself to make human clinical trials less necessary or cheaper).&#8221;</p><ul><li><p>it&#8217;s interesting to consider why this isn&#8217;t true for factors of production today. We live in a (relatively) capital-abundant and labor-scarce world. That is reflected in the labor share of income being 2x as high as the capital share of income. But this has been true for centuries upon centuries. Contra Piketty in &#8220;Capital in the 21st Century&#8221;, all these capital holders have not been able to get some runaway capital accumulation process going by figuring out a way around labor constraints. Why think that intelligence will be any different than capital in its ability to get around other factors of production? maybe the argument is that intelligence can actually help generate the other factors of production in a way that capital can&#8217;t.</p></li></ul></li></ul><h2><a href="https://sohl-dickstein.github.io/2024/02/12/fractal.html">Neural network training makes beautiful fractals</a> by Jascha Sohl-Dickstein</h2><p>Absolutely fascinating <a href="https://sohl-dickstein.github.io/2024/02/12/fractal.html">blog post</a>.</p><div id="vimeo-903855670" class="vimeo-wrap" data-attrs="{&quot;videoId&quot;:&quot;903855670&quot;,&quot;videoKey&quot;:&quot;&quot;,&quot;belowTheFold&quot;:true}" data-component-name="VimeoToDOM"><div class="vimeo-inner"><iframe src="https://player.vimeo.com/video/903855670?autoplay=0" frameborder="0" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" loading="lazy"></iframe></div></div><p>You want to train your model at the highest possible learning rate under which it still converges. But the boundary of convergence versus divergence is fractal, which makes these hyperparameters really hard to optimize for via gradient descent.</p><p>Now you can ask the question: evolution somehow found the right hyperparameters to train our brains. How did evolution solve this wicked problem? Presumably because gradient free optimization fares better against these kinds of fractal landscapes - if you optimize for the part of the region where the average speed of convergence is high (rather than just take the gradient from a specific point that&#8217;s bounded in an unpredictable way by fractals), seems like you could do much better.</p><p>Backing up, why is the meta-loss landscape fractal in the first place? Jascha&#8217;s explanation is that fractals often emerge when iteratively applying a function. Gradient descent on the parameters is one such function that you iterate across training steps. But then the follow up question is this. There&#8217;s lots of other iterative functions you could think of, even within the context of neural networks. Do they all lead to fractals? For example:</p><ul><li><p>In chain of thought, you apply a model to a string, which makes a new string, to which you apply the model, etc.</p></li><li><p>RNNs keep applying the same parameters to the hidden state.</p></li></ul><p>Over conversation, an AI researcher friend revealed that CoT and RNNs both have variance problems that could well be explained by these fractal like dynamics. Though I only understand this claim at a hand-wavy level.</p>]]></content:encoded></item><item><title><![CDATA[Thoughts on AI progress (Dec 2025)]]></title><description><![CDATA[Why I'm moderately bearish in the short term, and explosively bullish in the long term]]></description><link>https://www.dwarkesh.com/p/thoughts-on-ai-progress-dec-2025</link><guid isPermaLink="false">https://www.dwarkesh.com/p/thoughts-on-ai-progress-dec-2025</guid><dc:creator><![CDATA[Dwarkesh Patel]]></dc:creator><pubDate>Tue, 02 Dec 2025 21:39:14 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!QEPJ!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F90fa9666-5b8b-4685-a8fb-4b64cb7e0333_1080x1080.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h3>What are we scaling?</h3><p>I&#8217;m confused why some people have short timelines and at the same time are bullish on the current scale up of reinforcement learning atop LLMs. If we&#8217;re actually close to a human-like learner, this whole approach of training on verifiable outcomes is doomed.</p><p>Currently the labs are trying to bake in a bunch of skills into these models through &#8220;mid-training&#8221; - there&#8217;s an entire supply chain of companies building RL environments which teach the model how to navigate a web browser or <a href="https://fortune.com/2025/10/22/sam-altman-openai-wall-street-junior-bankers-ai-entry-level-jobs/">use Excel to write financial models</a>.</p><p>Either these models will soon learn on the job in a self directed way - making all this pre-baking pointless - or they won&#8217;t - which means AGI is not imminent. Humans don&#8217;t have to go through a special training phase where they need to rehearse every single piece of software they might ever need to use.</p><p>Beren Millidge made interesting points about this in a recent <a href="https://www.beren.io/2025-08-02-Most-Algorithmic-Progress-is-Data-Progress/">blog post</a>:</p><blockquote><p>When we see frontier models improving at various benchmarks we should think not just of increased scale and clever ML research ideas but billions of dollars spent paying PhDs, MDs, and other experts to write questions and provide example answers and reasoning targeting these precise capabilities ... In a way, this is like a large-scale reprise of the expert systems era, where instead of paying experts to directly program their thinking as code, they provide numerous examples of their reasoning and process formalized and tracked, and then we distill this into models through behavioural cloning. This has updated me slightly towards longer AI timelines since given we need such effort to design extremely high quality human trajectories and environments for frontier systems implies that they still lack the critical core of learning that an actual AGI must possess.</p></blockquote><p>You can see this tension most vividly in robotics. In some fundamental sense, robotics is an algorithms problem, not a hardware or data problem &#8212; with very little training, humans can learn how to teleoperate current hardware to do useful work. So if we had a human like learner, robotics would (in large part) be solved. But the fact that we don&#8217;t have such a learner makes it necessary to go out into a thousand different homes to learn how to pick up dishes or fold laundry.</p><p>One counterargument I&#8217;ve heard from the takeoff-within-5-years crew is that we have to do this cludgy RL in service of building a superhuman AI researcher, and then the million copies of automated Ilya can go figure out how to solve robust and efficient learning from experience.</p><p>This gives the vibes of that old joke, &#8220;We&#8217;re losing money on every sale, but we&#8217;ll make it up in volume.&#8221; Somehow this automated researcher is  going to figure out the algorithm for AGI - a problem humans have been banging their head against for the better part of a century - while not having the basic learning capabilities that children have? I find this super implausible.</p><p>Besides, even if you think the RLVR scaleup will soon help us automate AI research, the labs&#8217; actions suggest otherwise. You don&#8217;t need to pre-bake the consultant&#8217;s skills at crafting Powerpoint slides in order to automate Ilya. So clearly the labs&#8217; actions hint at a world view where these models will continue to fare poorly at generalizing and on-the-job learning, thus making it necessary to build in the skills that they hope will be economically valuable.</p><p>Another counterargument you could make is that even if the model could learn these skills on the job, it is just so much more efficient to build them up just once during training rather that again and again for each user or company. And look, it makes a lot of sense to just bake in fluency with common tools like browsers and terminals. Indeed one of the key advantages that AGIs will have is this greater capacity to share knowledge across copies. But people are underrating how much company and context specific skills are required to do most jobs. And there just isn&#8217;t currently a robust efficient way for AIs to pick up those skills.</p><h3>Human labor is valuable precisely because it&#8217;s not shleppy to train</h3><p>I was at a dinner with an AI researcher and a biologist. The biologist said she had long timelines. We asked what she thought AI would struggle with. She said her work has recently involved looking at slides and decide if a dot is actually a macrophage or just looks like one. The AI researcher says, &#8220;Image classification is a textbook deep learning problem&#8212;we could easily train for that.&#8221;</p><p>I thought this was a very interesting exchange, because it revealed a key crux between me and the people who expect transformative economic impacts in the next few years. Human workers are valuable precisely because we don&#8217;t need to build schleppy training loops for every small part of their job. It&#8217;s not net-productive to build a custom training pipeline to identify what macrophages look like given the way this particular lab prepares slides, then another for the next lab-specific micro-task, and so on. What you actually need is an AI that can learn from semantic feedback or from self directed experience, and then generalize, the way a human does.</p><p>Every day, you have to do a hundred things that require judgment, situational awareness, and skills &amp; context learned on the job. These tasks differ not just across different people, but from one day to the next even for the same person. It is not possible to automate even a single job by just baking in some predefined set of skills, let alone all the jobs.</p><p>In fact, I think people are really underestimating how big a deal actual AGI will be because they&#8217;re just imagining more of this current regime. They&#8217;re not thinking about billions of human-like intelligences on a server which can copy and merge all their learnings. And to be clear, I expect this (aka actual AGI) in the next decade or two. That&#8217;s fucking crazy!</p><h3>Economic diffusion lag is cope for missing capabilities</h3><p>Sometimes people will say that the reason that AIs aren&#8217;t more widely deployed across firms and already providing lots of value (outside of coding) is that technology takes a long time to diffuse. I think this is cope. People are using this cope to gloss over the fact that these models just lack the capabilities necessary for broad economic value.</p><p>Steven Byrnes has an <a href="https://www.lesswrong.com/posts/xJWBofhLQjf3KmRgg/four-ways-learning-econ-makes-people-dumber-re-future-ai">excellent post</a> on this and many other points:</p><blockquote><p>New technologies take a long time to integrate into the economy? Well ask yourself: how do highly-skilled, experienced, and entrepreneurial immigrant humans manage to integrate into the economy immediately? Once you&#8217;ve answered that question, note that AGI will be able to do those things too.</p></blockquote><p>If these models were actually like humans on a server, they&#8217;d diffuse incredibly quickly. In fact, they&#8217;d be so much easier to integrate and onboard than a normal human employee (they could read your entire Slack and Drive in minutes and immediately distill all the skills your other AI employees have). Plus, hiring is very much like a <a href="https://en.wikipedia.org/wiki/The_Market_for_Lemons">lemons market</a>, where it&#8217;s hard to tell who the good people are, and hiring someone bad is quite costly. This is a dynamic you wouldn&#8217;t have to worry about when you just wanna spin up another instance of a vetted AGI model.</p><p>For these reasons, I expect it&#8217;s going to be much much easier to diffuse AI labor into firms than it is to hire a person. And companies hire lots of people all the time. If the capabilities were actually at AGI level, people would be willing to spend trillions of dollars a year buying tokens (knowledge workers cumulatively earn 10s of trillions of dollars of wages a year). The reason that lab revenue are 4 orders of magnitude off right now is that the models are nowhere near as capable as human knowledge workers.</p><h3>Goal post shifting is justified</h3><p>AI bulls will often criticize AI bears for repeatedly moving the goal posts. This is often fair. AI has made a ton of progress in the last decade, and it&#8217;s easy to forget that.</p><p>But some amount of goal post shifting is justified. If you showed me Gemini 3 in 2020, I would have been certain that it could automate half of knowledge work. We keep solving what we thought were the sufficient bottlenecks to AGI (general understanding, few shot learning, reasoning), and yet we still don&#8217;t have AGI (defined as, say, being able to completely automate 95% of knowledge work jobs). What is the rational response?</p><p>It&#8217;s totally reasonable to look at this and say, &#8220;Oh actually there&#8217;s more to intelligence and labor than I previously realized. And while we&#8217;re really close to (and in many ways have surpassed) what I would have defined as AGI in the past, the fact that model companies are not making trillions is revenue clearly reveals that my previous definition of AGI was too narrow.&#8221;</p><p>I expect this to keep happening into the future. I expect that by 2030 that the labs will have made significant progress on my hobby horse of continual learning, and the models will start earning 100s of billions in revenue, but they won&#8217;t have automated all knowledge work, and I&#8217;ll be like, &#8220;We&#8217;ve made a lot of progress, but we&#8217;re not at AGI yet. We also need X, Y, and Z thing to get to trillions in revenue.&#8221;</p><p>Models keep getting more impressive at the rate the short timelines people predict, but more useful at the rate the long timelines people predict.</p><h3>RL scaling is laundering the prestige of pretraining scaling</h3><p>With pretraining, we had this extremely clean and general trend in improvement in loss across multiple orders of magnitude of compute (albeit on a power law, which is as weak as exponential growth is strong). People are trying to launder the presitge of pretraining scaling, which was almost as predictable as a physical law of the universe, to justify bullish projections about RLVR, for which we have no well fit publicly known trend. When intrepid researchers do try to piece together the implications from scarce public datapoints, they get quite bearish results. For example, Toby Ord has a <a href="https://www.tobyord.com/writing/how-well-does-rl-scale">great post</a> where he cleverly connects the dots between different o-series benchmark charts, which suggested &#8220;we need something like a 1,000,000x scale-up of total RL compute to give a boost similar to a GPT level&#8221;.</p><h3>Comparison to human distribution will make us at first overestimate (and then underestimate) AI</h3><p>There is huge variance in the amount of value that different humans can add, especially in white collar with its <a href="https://en.wikipedia.org/wiki/O-ring_theory_of_economic_development">O-ring dynamics</a>. The village idiot adds ~0 value to knowledge work, while top AI researchers are worth billions of dollars to Mark Zuckerberg.</p><p>AI models at any given snapshot of time, however, are roughly equally capable. Humans have all this variance, whereas AI models don&#8217;t. Because a disproportionate share of value-add in knowledge work comes from the top percentile humans, if we try to compare the intelligence of these AI models to the median human, then we will systematically overestimate the value they can generate. But by the same token, when models finally do match top human performance, their impact might be quite explosive.</p><h3>Broadly deployed intelligence explosion</h3><p>People have spent a lot of time talking about a software only singularity (where AI models write the code for a smarter successor system), a software + hardware singularity (where AIs also improve their successor&#8217;s computing hardware), or variations therein.</p><p>All these scenarios neglect what I think will be the main driver of further improvements atop AGI: continual learning. Again, think about how humans become more capable at anything. It&#8217;s mostly from experience in the relevant domain.</p><p>Over conversation, <a href="https://www.beren.io/">Beren Millidge</a> made the interesting suggestion that the future might look continual learning agents going out, doing jobs and generating value, and then bringing all their learnings back to the hive mind model, which does some kind of batch distillations on all these agents. The agents themselves could be quite specialized - containing what Karpathy called &#8220;the cognitive core&#8221; plus knowledge and skills relevant to the job they&#8217;re being deployed to do.</p><p>&#8220;Solving&#8221; continual learning won&#8217;t be a singular one-and-done achievement. Instead, it will feel like solving in context learning. GPT-3 demonstrated that in context learning could be very powerful (its ICL capabilities were so remarkable that the title of the GPT-3 <a href="https://arxiv.org/abs/2005.14165">paper</a> is &#8216;Language Models are Few-Shot Learners&#8217;). But of course, we didn&#8217;t &#8220;solve&#8221; in-context learning when GPT-3 came out - and indeed there&#8217;s plenty of progress still to be made, from comprehension to context length. I expect a similar progression with continual learning. Labs will probably release something next year which they call continual learning, and which will in fact count as progress towards continual learning. But human level continual learning may take another 5 to 10 years of further progress.</p><p>This is why I don&#8217;t expect some kind of runaway gains to the first model that cracks continual learning, thus getting more and more widely deployed and capable. If you had fully solved continual learning drop out of nowhere, then sure, it&#8217;s &#8220;game set match&#8221;, as Satya put it. But that&#8217;s not what&#8217;s going to happen. Instead, some lab is going to figure out how to get some initial traction on the problem. Playing around with this feature will make it clear how it was implemented, and the other labs will soon replicate this breakthrough and improve it slightly.</p><p>There&#8217;ll also probably be diminishing returns from learning-from-deployment. Each of the first 1000 consultant agents are each learning a ton from deployment. Less so the next 1000. And is there such a long tail to consultant work that the millionth deployed instance is likely to see something super important the other 999,999 instances missed? In fact, I wouldn&#8217;t be surprised if continual learning also ends up leading to a power law, but with respect to the number of instances deployed.</p><p>Besides, I just have some prior that competition will stay fierce, informed by the observation that all these previous supposed flywheels (user engagement on chat, synthetic data, etc) have done very little to diminish the greater and greater competition between model companies. Every month (or less), the big three will rotate around the podium, with other competitors not that far behind. There is some force (potentially talent poaching, rumor mills, or reverse engineering) which has so far neutralized any runaway advantages a single lab might have had.</p>]]></content:encoded></item><item><title><![CDATA[Podcast Strategy Doc (December 2025)]]></title><description><![CDATA[Back to The Lunar Society mission]]></description><link>https://www.dwarkesh.com/p/dec-strategy-doc</link><guid isPermaLink="false">https://www.dwarkesh.com/p/dec-strategy-doc</guid><dc:creator><![CDATA[Dwarkesh Patel]]></dc:creator><pubDate>Mon, 01 Dec 2025 18:35:34 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!QEPJ!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F90fa9666-5b8b-4685-a8fb-4b64cb7e0333_1080x1080.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h3>The mission</h3><p>I originally titled my podcast The Lunar Society. I changed it to Dwarkesh Podcast eventually because people kept thinking it was a crypto podcast (&#8221;to the moon!!!&#8221;). I named it after The Lunar Society of Birmingham, an informal club that met in the late 18th century. Members included James Watt, Matthew Boulton, Erasmus Darwin, Joseph Priestley, and Josiah Wedgwood. These were the scientists, inventors, and philosophers who had made first contact with the Industrial Revolution which was just starting to take shape around them. And they discussed everything from steam engines to abolition to chemistry to education reform.</p><p>Someday people will look back on this period the way we look back on the Enlightenment. Great thinkers having important debates right as the world was about to undergo these massive technological, economic, and political revolutions. And some of these thinkers actually managed to get a couple of the big things right.</p><p>Whatever happens next, I want the debates to have happened on this podcast, and to have happened well.</p><h3>We are moving from the age of podcasts to the age of essays</h3><p>I wanna make essays a first class citizen of what I do. This is for a couple of reasons:</p><ul><li><p>Interviews are best when I have some take that I can bounce against my guest. You only get to see Federer&#8217;s skill when he&#8217;s rallying against a decent player, and certainly not if he&#8217;s just bouncing the ball against a wall.</p></li><li><p>As AI becomes more and more closed off, the best people will not be in a position where they can explain their thinking clearly. This is why the Karpathy episode was so incredible. It&#8217;s rare to get an industry expert without any particular thing to pitch, and who can talk openly about the research. But I&#8217;m not aware of anyone else who is Karpathy-tier, and who is not obliged to keep his or her mouth shut about a couple of things.</p></li><li><p>My essays have done much better than my expectations, in terms of reach, correctness and impact. I wrote the continual learning essay on a whim one afternoon, because I wanted to articulate why all these LLM scripts I&#8217;ve written for my business haven&#8217;t been helpful. And I&#8217;m still a little shocked to realize that I had stumbled upon (at least part of) <a href="https://www.dwarkesh.com/p/ilya-sutskever-2?open=false#%C2%A7ssis-model-will-learn-from-deployment">what Ilya is working on at SSI</a>. It&#8217;s not a crazy insight by any means, but it&#8217;s notable that you can just think about stuff, and there&#8217;s a good chance you&#8217;ll figure out what&#8217;s up. Btw, after I released the essay, both <a href="https://dataconomy.com/2025/08/07/gpt-5-is-officially-out/#:~:text=He%20elaborated%2C%20%E2%80%9CThis%20is%20clearly,a%20model%20that%20continuously%20learns">Sam Altman</a> and <a href="https://x.com/The_AI_Investor/status/1967025620426387639">Demis Hassabis</a> have said that continual learning is a major bottleneck on the path to AGI. Of course, there&#8217;s no way to know whether they read my essay. But honestly, even if they hadn&#8217;t, I&#8217;d still be pretty stoked if I had independently pointed my finger at the exact same bottleneck as these guys, despite all their additional context.</p></li><li><p>Which brings to my next point. I feel like there&#8217;s actually not that many secrets. The researchers and CEOs of the AI labs are a couple months ahead of you. This just doesn&#8217;t amount to any substantial secret knowledge that, if only you knew, you&#8217;d also have 2027 timelines. A ton of progress has been made in the last 3 years since ChatGPT, but none of it was super shocking based on the rumor mill and some connecting of the dots. And then there&#8217;s the big picture questions about AI&#8217;s impacts, where your thinking might very plausibly be much better than people at the labs, just because it takes time to think, and these people are busy running a damn company.</p></li><li><p>Some of the questions I&#8217;m most interested in simply can&#8217;t be answered extemporaneously by any human being on the planet. They require knowledge across multiple different fields, and couple hours (to days) of crunching the numbers or thinking through shit.</p></li><li><p>Because often enough my guests can&#8217;t just answer pretty complicated fractal questions in a satisfying way on the spot, I get frustrated with the whole enterprise. The main angst I&#8217;ve kept receding back to over and over is, &#8220;Okay what did I actually learn from this interview? And if <em>I</em> didn&#8217;t get that much concrete insight and understanding out of it, despite a week+ of research and hours of conversation, what hope is there for the audience? And if no one learned anything, what the fuck are we doing here?&#8221; I feel much essays survive this cynicism much better. For example, I&#8217;m often frustrated that social scientists won&#8217;t speculate with me about what their insights imply about AI civilization, or historians about how history might have turned out differently given different counterfactuals. But it&#8217;s ridiculous to count on a scholar who is thinking about AGI for first time in his life to start shooting off some galaxy brain implications from his theory. But <em>I</em> can go read their books, and use my understanding of the technology to come up with some hot takes.</p></li><li><p>I can easily co-release my essays as narrations on my podcast and YouTube feed, so actually the essays are super complementary to this audio/video audience I&#8217;ve built up.</p></li></ul><h3>Gratitude</h3><p>In the spirit of Thanksgiving: a lottery winner who then won another lottery is less lucky than I am.</p><p>Every once in a while, I&#8217;ll be grabbing dinner with a writer whose work I was obsessed with in college. And a part of me is just like, &#8220;What the fuck is happening right now?&#8221; Many of my greatest intellectual heroes are now my direct friends and teachers. My job is to spend a week learning about whatever I&#8217;m most interested in, and then talk to the world expert on that topic. A job I would <em>pay</em> to do has rewarded me - intellectually, financially, socially - beyond my wildest expectations. And there&#8217;s millions of people who are into this stuff! This audience contains some of the smartest people in the world, including many of the people <em>I</em> am a huge fan of. Then there&#8217;s my team. It&#8217;s unreal how talented, agentic, tasteful, and detail-oriented my colleagues are. I genuinely have no idea how I convinced people this good to come run <em>a podcast</em>.</p>]]></content:encoded></item><item><title><![CDATA[RL is even more information inefficient than you thought]]></title><description><![CDATA[And implications for RLVR progress]]></description><link>https://www.dwarkesh.com/p/bits-per-sample</link><guid isPermaLink="false">https://www.dwarkesh.com/p/bits-per-sample</guid><dc:creator><![CDATA[Dwarkesh Patel]]></dc:creator><pubDate>Mon, 17 Nov 2025 16:54:09 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!J6SR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea6b8a9c-18d2-4a0f-a940-f04f83fcdd3c_989x690.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Recently, <a href="https://www.tobyord.com/writing/inefficiency-of-reinforcement-learning">people</a> have been <a href="https://thinkingmachines.ai/blog/lora/#how-much-capacity-is-needed-by-supervised-and-reinforcement-learning">talking</a> about how it takes way more FLOPs to get a single sample in RL than it does in supervised learning. In pretraining, you get a signal on every single token you train on. In RL, you have to unroll a whole thinking trajectory that&#8217;s 10s of 1000s of tokens long in order to get a single reward signal at the end (for example, did the unit test for my code pass/did I get the right answer to this math problem/etc).</p><p>But this is only half the problem.  Here&#8217;s a simple way to compare the learning efficiency of reinforcement learning versus supervised learning:</p><p>Bits/FLOP = Samples/Flop * Bits/Sample.</p><p>What I haven&#8217;t heard people talk about is the other term in our equation: Bits/Sample. And for most of training, the information density per sample is way way lower for RL.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.dwarkesh.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.dwarkesh.com/subscribe?"><span>Subscribe now</span></a></p><h3>Putting things in plain English</h3><p>In supervised learning (aka pretraining), you&#8217;re just soaking up bits. Every token is a hint at the structure of language, and the mind crafting that language, and the world that mind is seeing. Early in training, when you have a totally random model, you&#8217;re just maximally uncertain over all of this content. So each token is just blowing your mind. And you&#8217;re getting this exact signal of how wrong you were about the right answer, and what parameters you need to update to be less wrong.</p><p>Suppose you start with a randomly initialized model, and you kickstart training. If you&#8217;re doing next-token-prediction using supervised learning on &#8220;The sky is&#8221;, the training loop goes, &#8220;It&#8217;s actually &#8216;blue&#8217;. You said the probability of &#8216;blue&#8217; is .001%. Make the connections that were suggesting &#8216;blue&#8217; way way stronger. Alright, next token.&#8221;</p><p>In RL with policy gradient, you upweight all the trajectories where you get the answer right, and downweight all the trajectories where you get the answer wrong. But a model that&#8217;s not already very smart is just astonishingly unlikely to get the answer right.</p><p>If you were doing next-token-prediction on &#8220;The sky is&#8221; with RL, the training loop would be something like, &#8220;Okay, &#8216;halcyon&#8217; is wrong. Don&#8217;t do the thing that led to saying &#8216;halycon&#8217; &#8230; Okay &#8216;serendipity&#8217; is wrong &#8230;&#8221; Rinse and repeat this guesswork for somewhere around the number of tokens you have in your vocabulary (on the order of 100,000).</p><h3>The details</h3><p>Let&#8217;s think about how maximum bits/sample change as the pass rate (p) changes. Pass rate here means how likely you are to say the correct answer. To keep this simple, let&#8217;s say the answer is a token long. Then the pass rate when you have a totally untrained model is just 1/ (size of your vocabulary).</p><p>In supervised learning, you get told exactly what the right label is for each sample. The amount of new information you learn corresponds to how surprised you are to learn the correct answer - the lower your pass rate (aka prior probability of the correct answer), the more you learned from seeing the correct label. The basic formula for entropy tells us that you can learn -log(p) bits/sample from supervised learning.</p><p>In RL, you only get told whether you got the right answer or not. The amount of new information you can extract is bounded by how uncertain you are about this binary outcome. If you almost always pass (p &#8776; 1) or almost always fail (p &#8776; 0), each trial is very unlikely to surprise you. You&#8217;ll learn most when the probability of passing is like a coin toss (p &#8776; 0.5). The basic formula for the information content of a binary random variable tells us that you can learn at most Entropy(p) = -p log(p) - (1-p) log(1-p)<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> bits/sample from RL. </p><p>Okay let&#8217;s plot this.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!r_Hv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8137f76-7d2c-467f-9755-cc20e766bbad_989x690.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!r_Hv!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8137f76-7d2c-467f-9755-cc20e766bbad_989x690.png 424w, https://substackcdn.com/image/fetch/$s_!r_Hv!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8137f76-7d2c-467f-9755-cc20e766bbad_989x690.png 848w, https://substackcdn.com/image/fetch/$s_!r_Hv!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8137f76-7d2c-467f-9755-cc20e766bbad_989x690.png 1272w, https://substackcdn.com/image/fetch/$s_!r_Hv!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8137f76-7d2c-467f-9755-cc20e766bbad_989x690.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!r_Hv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8137f76-7d2c-467f-9755-cc20e766bbad_989x690.png" width="989" height="690" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b8137f76-7d2c-467f-9755-cc20e766bbad_989x690.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:690,&quot;width&quot;:989,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:53460,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.dwarkesh.com/i/179158054?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8137f76-7d2c-467f-9755-cc20e766bbad_989x690.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!r_Hv!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8137f76-7d2c-467f-9755-cc20e766bbad_989x690.png 424w, https://substackcdn.com/image/fetch/$s_!r_Hv!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8137f76-7d2c-467f-9755-cc20e766bbad_989x690.png 848w, https://substackcdn.com/image/fetch/$s_!r_Hv!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8137f76-7d2c-467f-9755-cc20e766bbad_989x690.png 1272w, https://substackcdn.com/image/fetch/$s_!r_Hv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8137f76-7d2c-467f-9755-cc20e766bbad_989x690.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Doesn&#8217;t look terrible. Yes, pretraining is much better for half of the pass rate range, but then RL is better for the other half. However, this graph is super misleading. Because what the power law (in scaling laws) implies is that you need an equivalent amount of compute to cross each order of magnitude improvement in the pass rate. If it took you X many FLOPs to go from 1/100,000 pass rate to 1/10,000, then it will take you X many FLOPs to go from 1/10,000 pass rate to 1/1,000. So, we should actually chart the pass rate on a log scale - again, to account for how each increment in the x-axis corresponds to the same number of FLOPs.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!qLHC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca871c67-c896-4f33-a5aa-2513d012363e_990x690.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!qLHC!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca871c67-c896-4f33-a5aa-2513d012363e_990x690.png 424w, https://substackcdn.com/image/fetch/$s_!qLHC!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca871c67-c896-4f33-a5aa-2513d012363e_990x690.png 848w, https://substackcdn.com/image/fetch/$s_!qLHC!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca871c67-c896-4f33-a5aa-2513d012363e_990x690.png 1272w, https://substackcdn.com/image/fetch/$s_!qLHC!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca871c67-c896-4f33-a5aa-2513d012363e_990x690.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!qLHC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca871c67-c896-4f33-a5aa-2513d012363e_990x690.png" width="990" height="690" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ca871c67-c896-4f33-a5aa-2513d012363e_990x690.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:690,&quot;width&quot;:990,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:51074,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.dwarkesh.com/i/179158054?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca871c67-c896-4f33-a5aa-2513d012363e_990x690.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!qLHC!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca871c67-c896-4f33-a5aa-2513d012363e_990x690.png 424w, https://substackcdn.com/image/fetch/$s_!qLHC!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca871c67-c896-4f33-a5aa-2513d012363e_990x690.png 848w, https://substackcdn.com/image/fetch/$s_!qLHC!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca871c67-c896-4f33-a5aa-2513d012363e_990x690.png 1272w, https://substackcdn.com/image/fetch/$s_!qLHC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca871c67-c896-4f33-a5aa-2513d012363e_990x690.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Oh boy, is that a sad picture. The regime where RL has comparable information density per sample to pre-training is this tiny slice at the very end of training, when you&#8217;ve got a pretty reasonable model anyways.</p><p>And again, I want to emphasize that this is totally separate from the point that getting a single sample from RL (aka unrolling a full trajectory before getting any signal) might take upwards of a million times more compute.</p><h3>It&#8217;s even worse than this - variance</h3><p>The situation for RL early in training is actually even worse than described above. When the pass rate is low, your gradient estimate is going to be incredibly noisy and unpredictable. Either you don&#8217;t sample the correct answer at all in your batch, in which you get almost no information. Or you do, and you get this giant spike. You&#8217;re getting jerked around, which is terrible for performant training.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a></p><p>Interestingly, pretraining has the exact inverse problem. There, variance is super high at the END of training. As pretraining progresses, you exhaust more and more of the reducible loss (things your model can actually learn about the data). What remains is mostly the irreducible loss. The irreducible loss is the intrinsic unpredictability of internet text.</p><p>How should the prompt, &#8220;Bob&#8217;s favorite color is&#8221; end?  Depends on Bob. There&#8217;s not some correct answer which your super smart model can actually get good at predicting. But your super smart model is still getting a gradient update on whatever random answer someone put on the internet. And this noise is drowning out the true signal that the couple of actually learnable tokens in the batch are giving you. I don&#8217;t know if this is accurate, but it seems like this explosion of variance at the end of pretraining is relevant to why batch sizes are increased as pretraining progresses.</p><h3>Getting to the Goldilocks zone in RL</h3><p>If RL works best in the regime where your pass rate is &gt;&gt;1%, then this raises the question, how can we construct the RL training to get (and keep) models in this learning flow state?</p><p>For example, we can think of pretraining AND inference scaling as increasing the pass rate during RL, allowing you to extract far more bits per sample.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!J6SR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea6b8a9c-18d2-4a0f-a940-f04f83fcdd3c_989x690.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!J6SR!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea6b8a9c-18d2-4a0f-a940-f04f83fcdd3c_989x690.png 424w, https://substackcdn.com/image/fetch/$s_!J6SR!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea6b8a9c-18d2-4a0f-a940-f04f83fcdd3c_989x690.png 848w, https://substackcdn.com/image/fetch/$s_!J6SR!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea6b8a9c-18d2-4a0f-a940-f04f83fcdd3c_989x690.png 1272w, https://substackcdn.com/image/fetch/$s_!J6SR!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea6b8a9c-18d2-4a0f-a940-f04f83fcdd3c_989x690.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!J6SR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea6b8a9c-18d2-4a0f-a940-f04f83fcdd3c_989x690.png" width="989" height="690" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ea6b8a9c-18d2-4a0f-a940-f04f83fcdd3c_989x690.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:690,&quot;width&quot;:989,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!J6SR!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea6b8a9c-18d2-4a0f-a940-f04f83fcdd3c_989x690.png 424w, https://substackcdn.com/image/fetch/$s_!J6SR!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea6b8a9c-18d2-4a0f-a940-f04f83fcdd3c_989x690.png 848w, https://substackcdn.com/image/fetch/$s_!J6SR!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea6b8a9c-18d2-4a0f-a940-f04f83fcdd3c_989x690.png 1272w, https://substackcdn.com/image/fetch/$s_!J6SR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea6b8a9c-18d2-4a0f-a940-f04f83fcdd3c_989x690.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>It&#8217;s been noted that <a href="https://arxiv.org/pdf/2012.03107">curriculum learning in not especially helpful for pretraining</a>, but <a href="https://arxiv.org/pdf/1707.05300">often essential for RL</a>. This makes total sense when you think about how RL is only getting meaningful bits per sample in this Goldilocks zone of pass rate, so you really want to order the learning such that the difficulty of challenges increases in tandem with the model&#8217;s intelligence.</p><p>Our pass rate framework also gives us good intuitions for why self play has been so productive in the history of RL. If you&#8217;re competing against a player who is almost as good as you, you are balancing around a 50% pass rate, which peaks out the bits you get from a random binary variable.</p><p>But self play is not the only way we can imagine of keeping pass rate high through training. Perhaps we can come up with some proxy evaluation which is much more dense. Density here can mean one of two things:</p><ol><li><p>Samples/FLOP density: You estimate the final reward using this proxy evaluation, but much earlier on in the episode, saving you the compute of unrolling the full trajectory. This is what a value function does.</p></li><li><p>Bits/Sample density: You come up with a proxy objective which is much easier to pass than the actual test under question. The simplest example I can think of is a process-reward model which says, &#8220;Hey, this rollout got the wrong answer, but I can see that its reasoning was on the right track at the start. So let&#8217;s up-weight those early tokens.&#8221;</p></li></ol><p>Section 4.2 of the<a href="https://arxiv.org/abs/2501.12948"> Deepseek R1 paper</a> why so far, it&#8217;s been hard to develop useful proxy objectives like this for LLMs.</p><h3>Fewer bits, sure, but very valuable bits</h3><p>To be fair to RL, while you may be learning far fewer Bits/FLOP in RL, the bits you learn are very important. They are not apples-to-apples comparable to the bits in pretraining. This is for two key reasons:</p><ol><li><p>Pre-training is teaching you what the data manifold of the internet looks like, which is only partially and indirectly related to, &#8220;How do I perform economically valuable tasks?&#8221; Whereas RL has the promise of giving you the good stuff directly.</p></li><li><p>Even if the pre-training corpus contains the instructions about how to accomplish a specific task, it does not have the thinking trace which teaches the model about how to correct its mistakes, or leverage its jagged and non-human repertoire of skills to accomplish the task.</p></li></ol><p>The rebuttal is that those bits are only available for a small fraction of the pass rate range (again, weighted on a log scale to account for how pass rate is trash for most of training).</p><p>By the way, now we can understand all these claims about how RLVR is <a href="https://arxiv.org/abs/2510.07364v3">only eliciting the capabilities already latent in the pretrained model</a>. Of course that&#8217;s the case. If the pretrained model didn&#8217;t have a high enough pass rate to begin with, then RL would have atrocious bits/sample, and thus not be able to learn at all. Move 37 is obviously one famous example where RL did teach a model a de-novo strategy. It&#8217;s worth noting that AlphaGo was trained on self play (see above re how self play increases pass rate), and that AlphaGo was surprisingly compute intensive <a href="https://epoch.ai/data/ai-models">for its time</a>.</p><h3>The jaggedness of RL</h3><p>People have pointed out that RLVR is empirically just leading models to associate a thought pattern to a problem type rather than instilling a more general policy of stepping back and thinking through the best approach.</p><p>Think about it. How is it possible that we have models which are world-class at coding competitions but at the same time leave extremely foreseeable bugs and technical debt all throughout the codebase?</p><p>What explains this weird jaggedness? Perhaps RLVR can&#8217;t distinguish trajectories that were generated from a more generalizable procedure vs just greedily matching the problem shape to some associated thought process.</p><p>When you&#8217;re doing policy gradient rollouts, this more complex general policy is extremely unlikely to be ever be sampled, whereas the simple heuristic policy does get sampled and grows in frequency until it reaches <a href="https://en.wikipedia.org/wiki/Fixation_(population_genetics)">fixation</a>. Meanwhile, the general policy recedes further and further from sight.</p><p>Then the question is, how do we build a short bridge between simple heuristic solutions and the more complex general strategy? And will that bridge just spontaneously emerge as time horizons expand, thus potentially requiring generalization?</p><p>My concern is that this general policy of stepping back and making tasteful judgements based on your understanding of the world will continue to be hard to spot-light using verifiable rewards, even on longer time horizon tasks. And so the solution to this jaggedness will require a more robust training procedure, not just scaling RLVR.</p><h3>Human learning</h3><p>Here we&#8217;re only talking about the bits/sample learned from model free RL - aka from some binary outcome at the end of an episode. But of course humans are obviously learning way more efficiently than this. Think about a repeat entrepreneur. We say that she has a ton of hard-won wisdom and experience. Very little of that learning comes from the one bit of outcome from her previous episode (whether the startup succeeded or not).</p><p>It&#8217;s not clear what the ML analog is for human learning from experience. Clearly, our observations and reflections update our world model (independent of the outcome at the end). And this is playing a very important role in our learning.</p><p>Maybe we shouldn&#8217;t be asking how we model free RL to &#8776;50% pass rate, so that can squeeze out a full drop of information from the outcome. Maybe we should be asking, how do humans wring out the buckets of information from the environment?</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.dwarkesh.com/p/bits-per-sample?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.dwarkesh.com/p/bits-per-sample?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Basically, this equation is saying, Information learned from a binary outcome = p(sample is correct) * (information gained when sample is correct) + p(sample is incorrect) * (information gained when sample is incorrect).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Thank you to Lukas Berglund for spotting that my previous exposition on this point was incorrect.</p></div></div>]]></content:encoded></item><item><title><![CDATA[Thoughts on the AI buildout]]></title><description><![CDATA[Fab CapEx overhang, 1 GW a week, China privileged in long timelines, and much else]]></description><link>https://www.dwarkesh.com/p/thoughts-on-the-ai-buildout</link><guid isPermaLink="false">https://www.dwarkesh.com/p/thoughts-on-the-ai-buildout</guid><dc:creator><![CDATA[Dwarkesh Patel]]></dc:creator><pubDate>Wed, 22 Oct 2025 17:58:39 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!hH1A!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Facf8778a-45d1-402f-b341-7a14efa2f18e_1600x1066.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Sam Altman <a href="https://blog.samaltman.com/abundant-intelligence">says</a> he wants to &#8220;create a factory that can produce a gigawatt of new AI infrastructure every week.&#8221;</p><p>What would it take to make this vision happen? Is it even physically feasible in the first place? What would it mean for different energy sources, upstream CAPEX in everything from fabs to gas turbine factories, and for US vs China competition?</p><p>These are not simple questions to answer. We wrote this blog post to teach ourselves more about them. We were surprised by some of the things we learned.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.dwarkesh.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.dwarkesh.com/subscribe?"><span>Subscribe now</span></a></p><h2>The fab CapEx overhang</h2><p><strong>With a single year of earnings in 2025, Nvidia could cover the last 3 years of TSMC&#8217;s ENTIRE CapEx.</strong></p><p>TSMC has done a total of $150B of CapEx over the last 5 years. This has gone towards many things, including building the entire 5nm and 3nm nodes (launched in 2020 and 2022 respectively) and the advanced packaging that Nvidia now uses to make datacenter chips. With only 20% of TSMC capacity<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a>, Nvidia has generated $100B in earnings.</p><p>Suppose TSMC nodes depreciate over 5 years - this is enormously conservative (newly built leading edge fabs are profitable for more than 5 years). That would mean that <strong>in 2025, NVIDIA will turn around $6B in depreciated TSMC Capex value into $200B in revenue.</strong></p><p><strong>Further up the supply chain, a single year of NVIDIA&#8217;s revenue almost matched the past 25 years of total R&amp;D and capex from the five largest semiconductor equipment companies combined, including ASML, Applied Materials, Tokyo Electron...</strong></p><p>We think this situation is best described as a &#8216;fab capex&#8217; overhang.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_3hh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4efa18a-d097-4ef7-a414-9103e2bac03e_1640x677.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_3hh!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4efa18a-d097-4ef7-a414-9103e2bac03e_1640x677.png 424w, https://substackcdn.com/image/fetch/$s_!_3hh!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4efa18a-d097-4ef7-a414-9103e2bac03e_1640x677.png 848w, https://substackcdn.com/image/fetch/$s_!_3hh!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4efa18a-d097-4ef7-a414-9103e2bac03e_1640x677.png 1272w, https://substackcdn.com/image/fetch/$s_!_3hh!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4efa18a-d097-4ef7-a414-9103e2bac03e_1640x677.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_3hh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4efa18a-d097-4ef7-a414-9103e2bac03e_1640x677.png" width="1456" height="601" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d4efa18a-d097-4ef7-a414-9103e2bac03e_1640x677.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:601,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:119162,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.dwarkesh.com/i/176848811?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4efa18a-d097-4ef7-a414-9103e2bac03e_1640x677.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!_3hh!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4efa18a-d097-4ef7-a414-9103e2bac03e_1640x677.png 424w, https://substackcdn.com/image/fetch/$s_!_3hh!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4efa18a-d097-4ef7-a414-9103e2bac03e_1640x677.png 848w, https://substackcdn.com/image/fetch/$s_!_3hh!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4efa18a-d097-4ef7-a414-9103e2bac03e_1640x677.png 1272w, https://substackcdn.com/image/fetch/$s_!_3hh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4efa18a-d097-4ef7-a414-9103e2bac03e_1640x677.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The reason we&#8217;re emphasizing this point is that if you were to naively speculate about what would be the first upstream component to constrain long term AI CapEx growth, you wouldn&#8217;t talk about copper wires or transformers - you&#8217;d start with the most complicated things that humans have ever made - which are the fabs that make semiconductors. We were stunned to learn that the cost to build these fabs pales in comparison to how much people are <em>already</em> willing to pay for AI hardware!</p><p>Nvidia could literally subsidize entire new fab nodes if they wanted to. We don&#8217;t think they will actually directly do this (or will they, wink wink, <a href="https://nvidianews.nvidia.com/news/nvidia-and-intel-to-develop-ai-infrastructure-and-personal-computing-products">Intel deal</a>) but this shows how much of a &#8216;fab capex&#8217; overhang there is.</p><h2>Upstream suppliers to datacenters will have to expand production</h2><p>For the last two decades, datacenter construction basically co-opted the power infrastructure left over from US deindustrialization. One person we talked to in the industry said that until recently, every single data center had a story. Google&#8217;s first operated data center was across a former aluminum plant. The hyperscalers are used to repurposing the power equipment from old steel mills and automotive factories.</p><p>This is honestly a compelling ode to capitalism. As soon as one sector became more relevant, America was quickly and efficiently able to co-opt the previous one&#8217;s carcass. But now we are in a different regime. Not only are hyperscalers building new data centers at a much bigger scale than before, they are building them from scratch, and competing for the same inputs with each other - not least of which is skilled labor.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!VbHG!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b58e0c1-a695-48d9-8adf-f49a934e7a0b_1600x1258.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!VbHG!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b58e0c1-a695-48d9-8adf-f49a934e7a0b_1600x1258.png 424w, https://substackcdn.com/image/fetch/$s_!VbHG!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b58e0c1-a695-48d9-8adf-f49a934e7a0b_1600x1258.png 848w, https://substackcdn.com/image/fetch/$s_!VbHG!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b58e0c1-a695-48d9-8adf-f49a934e7a0b_1600x1258.png 1272w, https://substackcdn.com/image/fetch/$s_!VbHG!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b58e0c1-a695-48d9-8adf-f49a934e7a0b_1600x1258.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!VbHG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b58e0c1-a695-48d9-8adf-f49a934e7a0b_1600x1258.png" width="1456" height="1145" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0b58e0c1-a695-48d9-8adf-f49a934e7a0b_1600x1258.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1145,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!VbHG!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b58e0c1-a695-48d9-8adf-f49a934e7a0b_1600x1258.png 424w, https://substackcdn.com/image/fetch/$s_!VbHG!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b58e0c1-a695-48d9-8adf-f49a934e7a0b_1600x1258.png 848w, https://substackcdn.com/image/fetch/$s_!VbHG!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b58e0c1-a695-48d9-8adf-f49a934e7a0b_1600x1258.png 1272w, https://substackcdn.com/image/fetch/$s_!VbHG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b58e0c1-a695-48d9-8adf-f49a934e7a0b_1600x1258.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Even <a href="https://www.mckinsey.com/featured-insights/week-in-charts/the-data-center-dividend">McKinsey</a> thinks these CapEx numbers are not crazy ($6.7T CapEx cumulative through 2030 implies around $2T in 2030).</figcaption></figure></div><p>For you to deploy $2 trillion of AI CapEx a year, someone else needs to be producing $2 trillion worth of all the other datacenter components, not just the chips. The people upstream to the datacenters - the companies producing everything from copper wire to turbines to transformers and switchgear - would need to build more factories.</p><p>The problem is that those factories have to be amortized over a many decade lifespan. At usual margins, those factories are only worth building if this AI demand lasts for 10-30 years.</p><p>The companies building those factories are not AGI pilled - they&#8217;re decades old industrials running low margin businesses which have been burned by many previous swings in demand. In the early 2000s, electricity demand seemed set to explode, so gas turbine manufactures like GE, Siemens, and others massively expanded manufacturing capacity. Then demand collapsed, leaving them with huge (almost bankrupting) overcapacity.</p><p>If there&#8217;s a financial overhang not just for fabs, but also for other datacenter components, could hyperscalers simply pay higher margins to accelerate capacity expansion? Especially given that chips are currently 60-70% of datacenter CapEx, the hyperscalers might just tell the companies which are building the other 30% not to worry about long run demand: &#8220;We&#8217;ll make you whole with just a couple of years of outrageous margins.&#8221;</p><p>The largest gas turbine manufacturers (<a href="https://www.wsj.com/market-data/quotes/GEV/financials?gaa_at=eafs&amp;gaa_n=AWEtsqdLIq5xmImL8Qj0n6YyHvojUG4FVcRzlYngNzg2tD84TFGQSHOIZ9oYUzlWEcM%3D&amp;gaa_ts=68f85923&amp;gaa_sig=Pmu1iX05V0g0UIb8qxSm2iAEQc--7rZ19bk2Nsvde9ggMIlX8_OVXoYM_Z8UHI1_MhqnQpWsurZCQhfWN12ZZw%3D%3D">GE Vernova</a>, <a href="https://www.wsj.com/market-data/quotes/XE/XETR/ENR/financials/annual/cash-flow">Siemens Energy</a>, and <a href="https://www.wsj.com/market-data/quotes/JP/XTKS/6503/financials/annual/cash-flow">Mitsubishi Electric</a>) are expecting to make $100B from gas turbines over the next 5 years. That <a href="https://www.eia.gov/electricity/generatorcosts/">corresponds</a> to around 100GW of generation capacity. These companies are doing a combined CapEx of around $5B/yr across all their divisions.</p><p>If the hyperscalers were willing to pay $200B instead of $100B for this 100GW of generation (for example, to incentivize faster delivery), they&#8217;d effectively be covering 20 years of these turbine manufacturers&#8217; entire CapEx (at current rates).</p><p>To build 100 GW of datacenters, the hyperscalers are going to have to invest many trillions of dollars anyways. Power generation is only about 7% of a datacenter&#8217;s cost. If natural gas ends up being the fastest way to scale generation, then doubling the cost of generation to 14% in order to make sure the power comes online fast enough might be easily worth it.</p><p>We think a similar dynamic is probably true for most of the components in a data center. <a href="https://marginalrevolution.com/marginalrevolution/2023/09/do-not-underrate-the-elasticity-of-supply.html">Do not underrate the elasticity of supply</a>.</p><h2>The coming labor bottleneck?</h2><p>Labor might actually end up being the most acute shortage - we can&#8217;t simply stamp out more workers (at least, not yet).</p><p>The 1.2 GW Stargate facility in Abilene has a workforce of over 5,000 people. Of course, there will be greater efficiencies as we scale this up, but naively that looks like 417,000 people to build 100 GW. And that&#8217;s on the low end of 2030 AI power consumption estimates. We&#8217;re gonna need stadiums full of electricians, heavy equipment operators, ironworkers, HVAC technicians,... you name it.</p><p>For reference, there&#8217;s <a href="https://www.bls.gov/ooh/construction-and-extraction/electricians.htm">800K electricians</a> and <a href="https://www.bls.gov/iag/tgs/iag23.htm#workforce">8 million construction workers</a> in the US. We hear that this labor pool is aging fast, but at least over the next few years, it seems like reallocation and big salary offers should be able to ameliorate the labor bottleneck.</p><h2>$400B+ ARR by end of decade is plausible</h2><p>Anthropic and OpenAI&#8217;s combined AI CapEx per year (being done indirectly, <a href="https://semianalysis.com/2025/09/03/amazons-ai-resurgence-aws-anthropics-multi-gigawatt-trainium-expansion/">mostly by Amazon and Microsoft in 2025</a>) seems to be around $100B<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>.</p><p>Revenues for OpenAI and Anthropic have been <a href="https://epoch.ai/data/ai-companies">3xing a year</a> for the past 2 years. Together, they are on track to earn $20B in 2025.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!VSzF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2622e95-a4b4-4e92-ace2-2279ee25c7a6_1920x1080.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!VSzF!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2622e95-a4b4-4e92-ace2-2279ee25c7a6_1920x1080.png 424w, https://substackcdn.com/image/fetch/$s_!VSzF!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2622e95-a4b4-4e92-ace2-2279ee25c7a6_1920x1080.png 848w, https://substackcdn.com/image/fetch/$s_!VSzF!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2622e95-a4b4-4e92-ace2-2279ee25c7a6_1920x1080.png 1272w, https://substackcdn.com/image/fetch/$s_!VSzF!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2622e95-a4b4-4e92-ace2-2279ee25c7a6_1920x1080.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!VSzF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2622e95-a4b4-4e92-ace2-2279ee25c7a6_1920x1080.png" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a2622e95-a4b4-4e92-ace2-2279ee25c7a6_1920x1080.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:144288,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.dwarkesh.com/i/176848811?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2622e95-a4b4-4e92-ace2-2279ee25c7a6_1920x1080.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!VSzF!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2622e95-a4b4-4e92-ace2-2279ee25c7a6_1920x1080.png 424w, https://substackcdn.com/image/fetch/$s_!VSzF!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2622e95-a4b4-4e92-ace2-2279ee25c7a6_1920x1080.png 848w, https://substackcdn.com/image/fetch/$s_!VSzF!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2622e95-a4b4-4e92-ace2-2279ee25c7a6_1920x1080.png 1272w, https://substackcdn.com/image/fetch/$s_!VSzF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2622e95-a4b4-4e92-ace2-2279ee25c7a6_1920x1080.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This means they&#8217;re spending 5 times as much on CapEx as they&#8217;re earning in revenue. This will probably change over time - <a href="https://pages.stern.nyu.edu/~adamodar/New_Home_Page/datafile/capex.html">more mature industries usually have CapEx </a><em><a href="https://pages.stern.nyu.edu/~adamodar/New_Home_Page/datafile/capex.html">less</a></em><a href="https://pages.stern.nyu.edu/~adamodar/New_Home_Page/datafile/capex.html"> than sales</a>. But AI is really fast growing, so it makes sense to keep investing more than you&#8217;re making right now.</p><p>Currently, <a href="https://www.wsj.com/tech/ai/tech-ai-spending-company-valuations-7b92104b">America&#8217;s AI CapEx is $400B/year</a>. For AI to not be a bubble in the short term, the datacenters currently being built right now need to generate $400B in revenue over their lifetime. Will they?</p><p>Google, Facebook, etc. have already shown us that if you can make a product which is modestly useful to billions of people, you can generate $100s of billions in revenue a year (Google+Meta make $400B/yr from ads alone).</p><p>OpenAI is approaching 1 billion, unmonetized free users, and we think a $12B to $100B revenue scale up is plausible just from their current products (e.g., see this vision: <a href="https://semianalysis.com/2025/08/13/gpt-5-ad-monetization-and-the-superapp/">GPT-5 Set the Stage for Ad Monetization and the SuperApp</a>). The question lies more in whether they can make a GPT-6 (or other products) in 3-5 years time that looks promising and economically useful enough to bring them into the $400B+ revenue range.</p><p>Of course, questions of revenue ultimately come down to your timelines. If AI truly lives up to its promise, then it&#8217;s in the reference class (at the very least) of white-collar wages, which are $10s of trillions of dollars a year.</p><p>Do you think that AI models will be able to do much of what a software engineer does by the end of a decade? If the <a href="https://en.wikipedia.org/wiki/Software_engineering_demographics">27M Software engineers worldwide</a> are all on super charged $1000/month AI agent plans that double their productivity (for 10-20% of their salary), that would be $324B revenue already.</p><h2>Lead times</h2><p>It takes about two years to build a new GW+ datacenter&#8212;and that&#8217;s before factoring in time to debug new chip generations. This means if you want to deploy 2 trillion in capex in 2030, you need to plan that out in 2028. At current trends in perf/watt and perf/$, $2T corresponds to roughly 66 GW of AI datacenter capacity.</p><p>For the last few years, hyperscalers and labs have consistently wanted more compute than they had previously made plans to develop. If the hyperscalers are planning for a 30% CAGR over the next 5 years and instead end up wanting to average 40% (hitting $2T capex in 2030) that&#8217;s a 20GW gap they&#8217;ll have to close in 2030 relative to the long term plans.</p><p>Elon (who didn&#8217;t even have an AI company during the time in which a traditional hyperscaler would have had to precommit to building capacity) solved his constraints by doing <a href="https://newsletter.semianalysis.com/p/xais-colossus-2-first-gigawatt-datacenter">insane things</a>. It&#8217;s unclear whether mere mortals can assemble 60 GW of greenfield capacity when they&#8217;d only planned for 30 or 40 GW.</p><p>If AI demand continues outpacing advance planning, there will be enormous pressure to compress datacenter construction timelines. The question is: what does this mean for energy sources and datacenter design? Some energy sources have much longer lead times than others, and some datacenter designs are more amenable to rapid deployment.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!1l8u!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7bfc8a34-c997-4e37-a867-3299ebec2fee_1600x1043.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!1l8u!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7bfc8a34-c997-4e37-a867-3299ebec2fee_1600x1043.png 424w, https://substackcdn.com/image/fetch/$s_!1l8u!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7bfc8a34-c997-4e37-a867-3299ebec2fee_1600x1043.png 848w, https://substackcdn.com/image/fetch/$s_!1l8u!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7bfc8a34-c997-4e37-a867-3299ebec2fee_1600x1043.png 1272w, https://substackcdn.com/image/fetch/$s_!1l8u!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7bfc8a34-c997-4e37-a867-3299ebec2fee_1600x1043.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!1l8u!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7bfc8a34-c997-4e37-a867-3299ebec2fee_1600x1043.png" width="1456" height="949" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7bfc8a34-c997-4e37-a867-3299ebec2fee_1600x1043.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:949,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!1l8u!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7bfc8a34-c997-4e37-a867-3299ebec2fee_1600x1043.png 424w, https://substackcdn.com/image/fetch/$s_!1l8u!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7bfc8a34-c997-4e37-a867-3299ebec2fee_1600x1043.png 848w, https://substackcdn.com/image/fetch/$s_!1l8u!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7bfc8a34-c997-4e37-a867-3299ebec2fee_1600x1043.png 1272w, https://substackcdn.com/image/fetch/$s_!1l8u!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7bfc8a34-c997-4e37-a867-3299ebec2fee_1600x1043.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Chips overwhelm everything else in the total cost of ownership of a datacenter. This is because 1. Chips are really expensive, and 2. The data center shell can be depreciated over 12-20 years, whereas the chips are fully depreciated (and have to be replaced) every 3 years.</p><p>So in terms of choosing an energy source, you can see very clearly what the order of priority should be.</p><ol><li><p>Lead times - Every month that the shell is not set up is a month that the chips (which are the overwhelming majority of your cost) aren&#8217;t being used.</p></li><li><p>Non chip CapEx - Much more expensive than electricity OpEx over a 3 year period.</p></li><li><p>Electricity OpEx</p></li></ol><p>So you can see why natural gas, for example, is much preferred over current nuclear reactors. Nuclear has extremely low op-ex, but has extremely long lead times and high CapEx. Natural gas may not be renewable, but you can just set up a couple dozen gas turbines next to the datacenter, and get your chips whirring fast.</p><p>Solar panels themselves are very cheap, but can be expensive to levelize across night + seasons. If you&#8217;re going pure solar, you have to build a bunch of overcapacity (4-7 GW solar capacity for 1 GW data center due to typical <a href="https://www.solarnplus.com/how-to-calculate-solar-power-plant-capacity-factor/">15-25% capacity factor</a>) and add a ton of batteries. Otherwise, you&#8217;re risking your expensive chips sitting idle during winter or at night.</p><p>Solar farms also have a massive land footprint, and require massive labor for installation &#8211; good luck hiring 30,000 people to lay out <a href="https://caseyhandmer.wordpress.com/2024/03/12/how-to-feed-the-ais/">20,000 acres of solar panels</a> and batteries across the desert in order to power a single 1 GW datacenter<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a>. The largest solar park in the world is currently the <a href="https://www.nytimes.com/2025/10/10/business/china-solar-tibetan-plateau.html">Gonghe Talatan Solar Park</a>. It generates enough electricity for around 3 GW of smoothed continuous power - but this requires 15 GW peak capacity - meaning 7.2 million solar panels - that&#8217;s the area of seven Manhattans.</p><p>Aside from lead times and costs, you could ask the question: which energy source is physically plentiful enough to supply this demand? The answer, it turns out, is literally all of them. The theoretical limits of any of the energy sources mentioned are orders of magnitude higher than what is needed for even the most explosive end of decade AI scenarios.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!bb-K!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70dcf1bc-673e-44e1-8bd6-3f6a1a6a46a4_1600x873.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!bb-K!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70dcf1bc-673e-44e1-8bd6-3f6a1a6a46a4_1600x873.png 424w, https://substackcdn.com/image/fetch/$s_!bb-K!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70dcf1bc-673e-44e1-8bd6-3f6a1a6a46a4_1600x873.png 848w, https://substackcdn.com/image/fetch/$s_!bb-K!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70dcf1bc-673e-44e1-8bd6-3f6a1a6a46a4_1600x873.png 1272w, https://substackcdn.com/image/fetch/$s_!bb-K!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70dcf1bc-673e-44e1-8bd6-3f6a1a6a46a4_1600x873.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!bb-K!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70dcf1bc-673e-44e1-8bd6-3f6a1a6a46a4_1600x873.png" width="1456" height="794" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/70dcf1bc-673e-44e1-8bd6-3f6a1a6a46a4_1600x873.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:794,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!bb-K!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70dcf1bc-673e-44e1-8bd6-3f6a1a6a46a4_1600x873.png 424w, https://substackcdn.com/image/fetch/$s_!bb-K!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70dcf1bc-673e-44e1-8bd6-3f6a1a6a46a4_1600x873.png 848w, https://substackcdn.com/image/fetch/$s_!bb-K!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70dcf1bc-673e-44e1-8bd6-3f6a1a6a46a4_1600x873.png 1272w, https://substackcdn.com/image/fetch/$s_!bb-K!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70dcf1bc-673e-44e1-8bd6-3f6a1a6a46a4_1600x873.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Off grid?</h2><p>A key question is whether datacenters will go &#8220;off-grid&#8221;&#8212;generating power on-site rather than connecting to the utility grid. Some of the largest datacenters are already doing this, e.g., Meta&#8217;s Orion or XAI&#8217;s Colossus.</p><p>Why would datacenters want to make power themselves rather than relying on the grid? They&#8217;re trying to get around interconnection delays. Connecting large new electricity sources to the grid now takes <a href="https://www.construction-physics.com/p/inside-the-interconnection-queue">over 5 years</a>.</p><p>For 20+ years, US electricity consumption has been either flat or growing slowly. Grid operators are now expecting huge increases in demand from AI datacenters, manufacturing reshoring, and electrification all happening simultaneously.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9opn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a004fbd-8c65-4184-967d-25fbe8b9c588_834x601.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9opn!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a004fbd-8c65-4184-967d-25fbe8b9c588_834x601.png 424w, https://substackcdn.com/image/fetch/$s_!9opn!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a004fbd-8c65-4184-967d-25fbe8b9c588_834x601.png 848w, https://substackcdn.com/image/fetch/$s_!9opn!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a004fbd-8c65-4184-967d-25fbe8b9c588_834x601.png 1272w, https://substackcdn.com/image/fetch/$s_!9opn!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a004fbd-8c65-4184-967d-25fbe8b9c588_834x601.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9opn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a004fbd-8c65-4184-967d-25fbe8b9c588_834x601.png" width="834" height="601" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4a004fbd-8c65-4184-967d-25fbe8b9c588_834x601.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:601,&quot;width&quot;:834,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!9opn!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a004fbd-8c65-4184-967d-25fbe8b9c588_834x601.png 424w, https://substackcdn.com/image/fetch/$s_!9opn!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a004fbd-8c65-4184-967d-25fbe8b9c588_834x601.png 848w, https://substackcdn.com/image/fetch/$s_!9opn!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a004fbd-8c65-4184-967d-25fbe8b9c588_834x601.png 1272w, https://substackcdn.com/image/fetch/$s_!9opn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a004fbd-8c65-4184-967d-25fbe8b9c588_834x601.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">ERCOT&#8217;s <a href="https://www.ercot.com/gridinfo/load/forecast">Annual Energy Forecast</a>. Who knew they were so AGI pilled?</figcaption></figure></div><p>One potential workaround: a <a href="https://nicholasinstitute.duke.edu/publications/rethinking-load-growth">Duke study </a>found that if datacenters agreed to curtail load just 0.25% of the time (roughly 22 hours per year), 76 GW of spare transmission capacity could be made available. Most transmission lines run well below capacity on average&#8212;the bottleneck only hits during peak demand.</p><p>But even if hyperscalers are able to perfectly capture this 76 GW, that only gets you from 2026-2028 in bullish AI scenarios. After that, either the grid expands or datacenters go off-grid.</p><h2>Distribution of datacenter sizes</h2><p>What will the distribution of individual datacenter sizes be? Here&#8217;s the argument for why we might end up seeing what looks like a thick sprinkle of 100 MW datacenters everywhere:</p><ul><li><p>If you can plop down a medium sized datacenter here and there, you can soak up any excess capacity in the grid. You can do this kind of arb with a 100 MW datacenter, but there&#8217;s no local excess capacity in the grid at the scale of 1 or 10 GW - that much power is on the scale of a whole grid itself!</p></li><li><p>For pretraining like learning, you want to have large contiguous blobs of compute. But already we&#8217;re moving to a regime of RL and midtraining, where learning involves a lot of inference. And the ultimate vision here is some kind of continual learning, where models are widely deployed through the economy and learning on the job/<a href="https://storage.googleapis.com/deepmind-media/Era-of-Experience%20/The%20Era%20of%20Experience%20Paper.pdf">from experience</a>. This seems compatible with medium sized datacenters housing 10s of thousands of instances of AIs working, generating revenue, and learning from deployment.</p></li></ul><p>Here&#8217;s the other vision. 1-10 GW datacenters, and then inference on device. Basically nothing in between.</p><ul><li><p>If we move to a world with vertically integrated industrial scale production of off-grid datacenters, maybe what you want to do is just buy a really big plot of land, build a big factory on site to stamp out as many individual compute halls and power/cooling/network blocks as possible. You can&#8217;t be bothered to build bespoke infrastructure for 100 MW here and there, when your company needs 50 GW total. A good analogy might be how a VC with billions to deploy won&#8217;t look at any deal smaller than deca millions.</p></li></ul><h2>A machine that spits out a GW a week</h2><p>Today&#8217;s datacenter construction resembles building a car in your driveway: the engine ships from Germany, transmission from Japan, harness from Detroit, and a mechanic spends months assembling it all on-site. Each datacenter is custom-built over 1-2 years, with networking, MEP systems, and racks assembled piece by piece.</p><p>You&#8217;re not getting to a <a href="https://blog.samaltman.com/abundant-intelligence">gigawatt per week</a> this way.</p><p>Could you have pre-fabricated compute halls? Fully wired racks, cooling systems, power equipment, batteries&#8212;assembled in factories and shipped as complete modules. Instead of 18 months of on-site construction, you&#8217;re sliding skids into place.</p><p>The design space is surprisingly flexible. Liquid cooling enables 500kW-1MW racks but requires totally different plumbing and construction. If solar dominates, maybe you go full DC-to-DC (panels generate DC, chips need DC) and skip all the AC conversion steps. Each design choice cascades into others&#8212;power source affects cooling approach affects rack density affects building design.</p><p>By the way, if AI turns out to be a bubble and we&#8217;re much further from AGI than Silicon Valley thinks, what lasting value gets built? You could tell a story about how the dot-com bubble paved the way for all the value the internet has generated. What&#8217;s the equivalent for this AI buildout?</p><p>The GPUs&#8212;70% of capex&#8212;are worthless after 3 years. The buildings and power infrastructure last decades but are overbuilt for non-AI workloads. Perhaps the enduring value is this new industrial capability: the ability to rapidly manufacture and deploy massive compute infrastructure on demand. Like how the dot-com bubble left us with fiber in the ground, maybe an AI bubble leaves us with an industrialized datacenter supply chain and an expanded electric grid.</p><p>Crypto actually did this for the current wave of AI. For example, Crusoe - which is helping OpenAI build out Stargate in Abilene - was founded to build Bitcoin mining datacenters on stranded natural gas.</p><h2>Does China win by default on long timelines?</h2><p>Why doesn&#8217;t China just win by default? For every component other than chips which is required for this industrial scale ramp up (solar panels, HV transformers, switchgear, new grid capacity), China is the dominant global manufacturer. China produces 1 TW of solar PV a year, whereas the US produces 20 GW (and even for those, the cells and wafers themselves are manufactured in China, and only the final module is assembled in the US).</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9S5W!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef11b1fc-18ee-4f5f-a2a0-187a89691daa_1600x1385.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9S5W!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef11b1fc-18ee-4f5f-a2a0-187a89691daa_1600x1385.png 424w, https://substackcdn.com/image/fetch/$s_!9S5W!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef11b1fc-18ee-4f5f-a2a0-187a89691daa_1600x1385.png 848w, https://substackcdn.com/image/fetch/$s_!9S5W!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef11b1fc-18ee-4f5f-a2a0-187a89691daa_1600x1385.png 1272w, https://substackcdn.com/image/fetch/$s_!9S5W!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef11b1fc-18ee-4f5f-a2a0-187a89691daa_1600x1385.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9S5W!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef11b1fc-18ee-4f5f-a2a0-187a89691daa_1600x1385.png" width="1456" height="1260" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ef11b1fc-18ee-4f5f-a2a0-187a89691daa_1600x1385.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1260,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!9S5W!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef11b1fc-18ee-4f5f-a2a0-187a89691daa_1600x1385.png 424w, https://substackcdn.com/image/fetch/$s_!9S5W!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef11b1fc-18ee-4f5f-a2a0-187a89691daa_1600x1385.png 848w, https://substackcdn.com/image/fetch/$s_!9S5W!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef11b1fc-18ee-4f5f-a2a0-187a89691daa_1600x1385.png 1272w, https://substackcdn.com/image/fetch/$s_!9S5W!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef11b1fc-18ee-4f5f-a2a0-187a89691daa_1600x1385.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Not only does China generate more than twice the electricity than the US, but that generation has been growing more than 10 times faster than in the US. The reason this is significant is that the power build out can be directed to new datacenter sites. China State Grid could collaborate with Alibaba, Tencent, and Baidu to build capacity where it is most helpful to the AI buildout, and avoid the zero-sum race in the US between different hyperscalers to take over capacity that already exists.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!HiMu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1d8146e-7798-43b8-a978-6a80435ea1dd_1473x653.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!HiMu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1d8146e-7798-43b8-a978-6a80435ea1dd_1473x653.png 424w, https://substackcdn.com/image/fetch/$s_!HiMu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1d8146e-7798-43b8-a978-6a80435ea1dd_1473x653.png 848w, https://substackcdn.com/image/fetch/$s_!HiMu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1d8146e-7798-43b8-a978-6a80435ea1dd_1473x653.png 1272w, https://substackcdn.com/image/fetch/$s_!HiMu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1d8146e-7798-43b8-a978-6a80435ea1dd_1473x653.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!HiMu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1d8146e-7798-43b8-a978-6a80435ea1dd_1473x653.png" width="1456" height="645" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c1d8146e-7798-43b8-a978-6a80435ea1dd_1473x653.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:645,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!HiMu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1d8146e-7798-43b8-a978-6a80435ea1dd_1473x653.png 424w, https://substackcdn.com/image/fetch/$s_!HiMu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1d8146e-7798-43b8-a978-6a80435ea1dd_1473x653.png 848w, https://substackcdn.com/image/fetch/$s_!HiMu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1d8146e-7798-43b8-a978-6a80435ea1dd_1473x653.png 1272w, https://substackcdn.com/image/fetch/$s_!HiMu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc1d8146e-7798-43b8-a978-6a80435ea1dd_1473x653.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Is China privileged in the long timelines world? SMIC will probably *eventually* catch up to TSMC (and maybe SMEE or SiCarrier to ASML, CXMT and YMTC to SK Hynix and Micron, NAURA and AMEC to Applied Materials, LAM, Tokyo Electron and KLA) - export controls won&#8217;t preserve the lead forever. If there&#8217;s not a software only intelligence explosion before 2030, and AI just becomes a massive industrial race across the entire supply chain from robotics to solar panels and batteries to steel, then why doesn&#8217;t China end up leading? Isn&#8217;t China&#8217;s differential advantage precisely these kinds of rapid and massive infrastructure build-outs?</p><p>Semianalysis <a href="https://semianalysis.com/2025/09/08/huawei-ascend-production-ramp/#china%e2%80%99s-compute-compared-to-americas">projects</a> that China will actually be able to ship fewer chips next year than it did this year, mainly because of domestic HBM production constraints. But we wonder how much that matters in the long to medium term.</p><p>GPUs are fully depreciated over 3 years (because new designs and better underlying process nodes make previous generations irrelevant). And lead times on building new datacenters are only a year or two long - datacenter design as a whole might be redone to accommodate industrial scale vertical integration.</p><p>All this makes us wonder whether every 3 years, the AI race starts totally fresh. While we might be able to constrain China&#8217;s production up till 2028 (or maybe even into the end of the decade with the lithography chokepoint), what&#8217;s the story for why this matters 2030 onwards?</p><h2>Two scenarios - AI winter, and AI explosion</h2><p>In order to get a handle on questions like these, we think it&#8217;s really helpful to just put numbers into a spreadsheet. Even if the scenarios are pulled out of your ass, they give you some sense of what would have to be true about the world in which they came to pass. For example, it&#8217;s interesting to ask: given trends in hardware price and performance, how much power would $2T CapEx correspond to in 2030? How about $500B?</p><p>We decided to chart two different potential trajectories:</p><ol><li><p>Explosive growth, where AI investment grows smoothly and then starts to accelerate on the back of booming economic growth from AI automation (30% GDP growth in 2035).</p></li><li><p>AI winter, where Investment growth crashes around 2029 and then grows at a smooth 5% annual growth rate from 2032 onwards.</p></li></ol><p>Here&#8217;s AI CapEx and AI power up to 2040 in both scenarios<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a>:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!hH1A!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Facf8778a-45d1-402f-b341-7a14efa2f18e_1600x1066.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!hH1A!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Facf8778a-45d1-402f-b341-7a14efa2f18e_1600x1066.png 424w, https://substackcdn.com/image/fetch/$s_!hH1A!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Facf8778a-45d1-402f-b341-7a14efa2f18e_1600x1066.png 848w, https://substackcdn.com/image/fetch/$s_!hH1A!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Facf8778a-45d1-402f-b341-7a14efa2f18e_1600x1066.png 1272w, https://substackcdn.com/image/fetch/$s_!hH1A!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Facf8778a-45d1-402f-b341-7a14efa2f18e_1600x1066.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!hH1A!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Facf8778a-45d1-402f-b341-7a14efa2f18e_1600x1066.png" width="1456" height="970" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/acf8778a-45d1-402f-b341-7a14efa2f18e_1600x1066.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:970,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!hH1A!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Facf8778a-45d1-402f-b341-7a14efa2f18e_1600x1066.png 424w, https://substackcdn.com/image/fetch/$s_!hH1A!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Facf8778a-45d1-402f-b341-7a14efa2f18e_1600x1066.png 848w, https://substackcdn.com/image/fetch/$s_!hH1A!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Facf8778a-45d1-402f-b341-7a14efa2f18e_1600x1066.png 1272w, https://substackcdn.com/image/fetch/$s_!hH1A!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Facf8778a-45d1-402f-b341-7a14efa2f18e_1600x1066.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!UR2B!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c13c47e-80a8-4951-9c9e-e63ca011482d_1600x1012.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!UR2B!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c13c47e-80a8-4951-9c9e-e63ca011482d_1600x1012.png 424w, https://substackcdn.com/image/fetch/$s_!UR2B!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c13c47e-80a8-4951-9c9e-e63ca011482d_1600x1012.png 848w, https://substackcdn.com/image/fetch/$s_!UR2B!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c13c47e-80a8-4951-9c9e-e63ca011482d_1600x1012.png 1272w, https://substackcdn.com/image/fetch/$s_!UR2B!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c13c47e-80a8-4951-9c9e-e63ca011482d_1600x1012.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!UR2B!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c13c47e-80a8-4951-9c9e-e63ca011482d_1600x1012.png" width="1456" height="921" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4c13c47e-80a8-4951-9c9e-e63ca011482d_1600x1012.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:921,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!UR2B!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c13c47e-80a8-4951-9c9e-e63ca011482d_1600x1012.png 424w, https://substackcdn.com/image/fetch/$s_!UR2B!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c13c47e-80a8-4951-9c9e-e63ca011482d_1600x1012.png 848w, https://substackcdn.com/image/fetch/$s_!UR2B!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c13c47e-80a8-4951-9c9e-e63ca011482d_1600x1012.png 1272w, https://substackcdn.com/image/fetch/$s_!UR2B!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c13c47e-80a8-4951-9c9e-e63ca011482d_1600x1012.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In the explosive growth scenario, Sam Altman&#8217;s vision of <a href="https://blog.samaltman.com/abundant-intelligence">1 GW a week</a> for the leading company comes true in 2036. But in that world, global AI power draw would be twice US&#8217;s current electricity generation.</p><p>We had a lot of fun building and playing around the spreadsheet that led to these scenario projections. <a href="https://docs.google.com/spreadsheets/d/1UDmU-4XenC4WJl2PpOkeU3L1uX9wKyrPiD3k6op-P3w/edit?usp=sharing">You might too</a>.</p><h2>Concluding thoughts</h2><p>Romeo and I decided to write this blog post because we had lots of questions about the AI buildout. We&#8217;re left with many more questions.</p><p>We don&#8217;t have any hard conclusions. What we&#8217;ve managed to do is assemble some considerations relevant to answering our questions.</p><p>There is no one person in the world who is the expert in all of the relevant fields. And even if there was such a person, they wouldn&#8217;t have a crystal ball. So we&#8217;ve spent a lot of time hoping on calls, reading PDFs, and talking to LLMs.</p><p>If you do have more information or thoughts, please reach out to us. We&#8217;re eager to learn more.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.dwarkesh.com/p/thoughts-on-the-ai-buildout?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.dwarkesh.com/p/thoughts-on-the-ai-buildout?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><h3>About the authors</h3><ul><li><p><a href="https://www.iaps.ai/romeo-dean">Romeo</a> is a researcher at AI Futures Project working on writing AI scenarios and recently graduated from Harvard. For their <a href="https://ai-2027.com/">first scenario</a> he focused on <a href="https://ai-2027.com/research/compute-forecast">compute forecasts</a> &#8211; now he&#8217;s also thinking about broader economic impacts for a new scenario with longer timelines and a positive vision for governance. If you have expertise in economics and an interest in future AI scenarios, please reach out at romeo@ai-futures.org.</p></li><li><p>Dwarkesh is a <a href="https://www.youtube.com/c/DwarkeshPatel">podcaster</a>. He&#8217;s keen to interview someone really good on all these topics. Reach out at hello@dwarkeshpatel.com.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.dwarkesh.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.dwarkesh.com/subscribe?"><span>Subscribe now</span></a></p></li></ul><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>In 2023 and 2024 Nvidia had grown to 11-12% of TSMC&#8217;s revenue (<a href="https://techsoda.substack.com/p/explainer-tsmcs-2024-annual-report">source 1</a>, <a href="https://www.tomshardware.com/tech-industry/analyst-estimates-nvidia-is-now-tsmcs-second-largest-customer-accounting-for-11-of-revenue-in-2023">source 2</a>). This year TSMC expects $110B total revenue, 26% higher than last year&#8217;s $88B (breaking from the 8% average increase the previous 2 years). Assuming this increase (26% vs. 8%) is from AI revenue, and in 2024 AI was already 15% of their revenue, that gives us an estimate of around $30B TSMC 2025 AI Revenue.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>OpenAI and Anthropic look on track to get access to around 3 GW of <a href="https://newsletter.semianalysis.com/p/xais-colossus-2-first-gigawatt-datacenter">combined AI capacity</a> this year (the exact number is obscure, you&#8217;d have to buy Semianalysis&#8217;s model to know exactly). At around $30/W of chip-only CapEx ($3.9M per 132kW NVL72 GB200 rack) that&#8217;s about $100B.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>These figures assume solar tracking systems that rotate panels to follow the sun throughout the day. Austin Vernon points out that fixed-tilt systems (panels just mounted on the ground) can be packed more densely and installed much faster&#8212;potentially requiring only ~6,000 acres and a few thousand workers versus 20,000 acres and 30,000 workers. The tradeoff is lower capacity factor, but you can compensate by installing more panels. Panels are cheap&#8212;installation time is expensive.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>For these scenarios, we&#8217;re assuming Moore&#8217;s Law continues, and that there&#8217;s continued improvements in power efficiency and cost efficiency of chips (albeit all three trends proceed at rates slightly slower than historically). Check out <a href="https://docs.google.com/spreadsheets/d/1UDmU-4XenC4WJl2PpOkeU3L1uX9wKyrPiD3k6op-P3w/edit?usp=sharing">the spreadsheet</a> for a full accounting of assumptions.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[On The Vital Question by Nick Lane]]></title><description><![CDATA[Life is continuous with Earth's geochemistry??]]></description><link>https://www.dwarkesh.com/p/the-vital-question</link><guid isPermaLink="false">https://www.dwarkesh.com/p/the-vital-question</guid><pubDate>Wed, 01 Oct 2025 14:03:06 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!GfM9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc299e01f-5d83-4b04-b903-e217885501be_804x548.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I just interviewed Nick Lane yesterday. It turned out great. I&#8217;m planning on publishing the episode Friday. </p><p>In the meantime, I am sharing the notes I was taking as I try to comprehend his book, <em><a href="https://www.amazon.com/Vital-Question-Evolution-Origins-Complex/dp/0393088812">The Vital Question: Energy, Evolution, and the Origins of Complex Life</a></em>. </p><p>The story the book tells of life&#8217;s evolution is both fascinating and super complicated. Hopefully there notes are helpful to others who are interested in this topic.</p><p>I wrote up these notes piecemeal, part by part, on Twitter, rather than all at once at the end. This blog compiles all of them in place.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.dwarkesh.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe for future book notes like this, and for the episode when it comes out.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>Part 1 - Why eukaryotes are so special</h2><p>In the intro he lists out the motivating questions: </p><p>Why are bacteria so relatively simple despite being around for 4 billion years? Why is there so much shared structure between all eukaryotic cells despite the enormous morphological variety between animals, plants, fungi, and protists? Why did the endosymbiosis event that led to eukaryotes happen only once, and in the particular way that it did? And why is all life powered by proton gradients? </p><p>Nick says all these questions are connected. </p><p>Lane says there&#8217;s 2 different philosophies on what bottlenecks evolutionary exploration: the niches made available by the environment, OR the internal structure necessary to exploit those niches. </p><p>Textbook view is that the environment constrains exploration, whereas structure is flexible and can accommodate once the right environment is in place. Nick Lane thinks it&#8217;s the opposite. </p><p>There&#8217;s been 2 big oxidation events - the first one (2.4 billion years ago) paved the way for eukaryotic cells. The second one (600 million years ago) led to the Cambrian explosion, resulting in all the variety in animals and plants and other complex life we see. So it seems the environment is central. Once you get a bunch of oxygen up in the air and into the oceans, you can start making all kinds of cool shit. </p><p>But hold on. Here&#8217;s what you&#8217;d expect to see if the environment was the key constraint: With this key unlock of aerobic respiration, different brands of bacteria independently evolve towards greater complexity to fill the new niches opened up (one masters osmotrophy and branches off into fungi, another photosynthesis, another phagocytosis, etc). However, you don&#8217;t see this. </p><p>Instead you see that all complex life emerges from a single common eukaryotic ancestor (2.2 billion years ago). There is no independent convergent evolution towards this kind of complexity (bacteria have had 4 billion years to evolve this kind of complexity, and have stayed remarkably similar through the whole time). </p><p>In fact, once you do get this key structural unlock, eukaryotic organisms proliferate widely, filling niches ranging from 100 feet long blue whales to 0.8 meter long picoplankton. </p><p>What&#8217;s more:</p><ul><li><p>The amount of shared structure between all eukaryotic cells is remarkable. They have almost all the same organelles and components. Nick writes: </p><p>&#8220;Most of us couldn&#8217;t distinguish between a plant cell, a kidney cell and a protist from the local pond down the electron microscope.&#8221;</p></li><li><p>There&#8217;s no intermediate proto-eukaryotes, which have some, but not all, of the functionality available to eukaryotic cells. This is wild given how evolution works. We have an extensive record of the incremental upgrades between photoreceptive amoebas and mammalian eyes. Why don&#8217;t we have proto-eukaryotic cells which reproduce via meiosis but don&#8217;t have compartmentalized nucleuses, or have mitochondria but no cytoskeleton? </p></li></ul><p>Nick argues that the fact that no such subset of eukaryotic traits exists suggests that it is not structurally possible to survive with only some fraction of eukaryotic equipment - you need the whole package all at once. </p><p>Obviously this raised the question of how the whole package was evolved at once. Which I think he will address in future chapters. </p><p>Some questions for Nick:</p><ul><li><p>If his view is that structure was the main bottleneck, and we&#8217;ve had eukaryotes for 2.2 billion years, then why didn&#8217;t we have all these animals and shit for 2 billion years? Why did they only arise 600 million years ago (aka the Cambrian explosion)?</p></li><li><p>Nick argues that eukaryotic cells are a much more significant unlock than multi-cellularity. Multi-cellularity evolved independently dozens of times, but we only have evidence of one event like the emergence of the first eukaryotic cell. If multi-cellularity evolved independently so many times (between fungi, slime molds, algae, etc etc), do we see interesting differences based on the situations in which they evolved? Do they regulate the differentiation of cells, the organization of the body differently, and communication between tissues differently? TODO look it up later. </p></li></ul><p>A tangential thought. This whole debate about whether structure or environment matters more seems analogous to the discussion in ML of whether architecture or data matters more. And there it seems like data is quite crucial, but for meta-learning and generality to kick off, the architecture has to make it possible for information to flow in the right way. For example, in context learning is a kind of meta-learning that arises only once the model has the capability to attend to hundreds of previous tokens, which became tractable with transformers.</p><h2>Part 2 - How the first cells evolved</h2><p>His main argument here is that life is continuous with the planet&#8217;s geochemistry. </p><p>Aka a lot of the main characteristics of cells - membranes, enzymes, energy via proton gradients - descend from spontaneous processes in the Earth. </p><p>But you can&#8217;t have these characteristics evolve piecemeal in different locations. You need one location that houses all the processes which could then give rise to the first cell. </p><p>Important context, by the way, is that all life descends from a single common ancestor - LUCA (last universal common ancestor). </p><p>Okay, so what candidate environment could give rise to LUCA? It needs two main characteristics:</p><ol><li><p>There&#8217;s a continuous flux of carbon and energy (in some sense, all life is a flux of carbon and energy, but you need some geochemistry to maintain this disequilibrium before the first cells can co-opt it).</p></li><li><p>Something which concentrates and catalyzes the reactions which lead to organics (aka inorganic equivalents of cells and enzymes). </p></li></ol><p>This rules out a lot of old theories: a warm pond with ammonia and salts and the odd lightning bolt doesn&#8217;t drive continuous flux, nor concentrate early organics in a cell-like volume to drive forward reactions. </p><p>Nick thinks alkaline sea vents are a unique fit to this challenge, and also help explain a lot of the contingent biochemistry that all life ended up using because of our shared inheritance. </p><p>Okay, let&#8217;s dig in: and for context, basically Nick here is trying to explain how you end up with an early version of the reverse Krebs cycle spontaneously. Reverse Krebs cycle takes in H2 and CO2 and makes organic molecules that are the precursors of fatty acids, proteins, and sugars. </p><p>Another important bit of context: All life runs on proton gradients. Burning food with oxygen (or other oxidants in anaerobic respiration) pumps H+ ions across a membrane, like filling a dam. These ions flow back through ATP synthase&#8212;a molecular turbine&#8212;which harnesses the flow to attach phosphate to ADP, creating ATP. Your body contains just 60 grams of ATP, but the ATP&#8594;ADP&#8594;ATP cycle is so rapid you process your body weight in ATP daily. </p><p>Sidenote: If a solution is acidic, it means there&#8217;s a lot of H+ ions in it. And if it&#8217;s basic (aka alkaline), it means there&#8217;s a lot of OH- ions in it. </p><p>Okay so what was happening in these alkaline hydrothermal vents? There&#8217;s 3 sides to this picture: the inside of the vent, the vent wall, and the ocean side of the vent.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!GfM9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc299e01f-5d83-4b04-b903-e217885501be_804x548.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!GfM9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc299e01f-5d83-4b04-b903-e217885501be_804x548.png 424w, https://substackcdn.com/image/fetch/$s_!GfM9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc299e01f-5d83-4b04-b903-e217885501be_804x548.png 848w, https://substackcdn.com/image/fetch/$s_!GfM9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc299e01f-5d83-4b04-b903-e217885501be_804x548.png 1272w, https://substackcdn.com/image/fetch/$s_!GfM9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc299e01f-5d83-4b04-b903-e217885501be_804x548.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!GfM9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc299e01f-5d83-4b04-b903-e217885501be_804x548.png" width="804" height="548" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c299e01f-5d83-4b04-b903-e217885501be_804x548.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:548,&quot;width&quot;:804,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:553258,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.dwarkesh.com/i/174939358?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc299e01f-5d83-4b04-b903-e217885501be_804x548.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!GfM9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc299e01f-5d83-4b04-b903-e217885501be_804x548.png 424w, https://substackcdn.com/image/fetch/$s_!GfM9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc299e01f-5d83-4b04-b903-e217885501be_804x548.png 848w, https://substackcdn.com/image/fetch/$s_!GfM9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc299e01f-5d83-4b04-b903-e217885501be_804x548.png 1272w, https://substackcdn.com/image/fetch/$s_!GfM9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc299e01f-5d83-4b04-b903-e217885501be_804x548.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>On the inside of the vent, you&#8217;ve got iron rich rock basically rusting, which lets out H2 and OH- into the stream of water piping through (aka making the water basic/alkaline). </p><p>The wall is made up of catalytic minerals like FeS, and also has a ton of tiny pores which connect the inside to the outside. </p><p>And the ocean side has a bunch of dissolved CO2 - early Earth was basically a giant ocean, but also had a lot of volcanoes that let out lots of CO2. And the oceans are quite acidic too, because CO2 becomes carbonic acid when dissolved in water. </p><p>Within the tiny pores inside these vents, you have H2 reacting with CO2 to form simple organics like formaldehyde (CH2O) and methanol (CH3OH), instigated by the FeS in the walls, which acts as a catalyst for this reaction. </p><p>Remedial chemistry: feel free to skip this para - I&#8217;m just going to include it since it took me some effort to relearn the high school chemistry involved. And it was quite satisfying to understand. Why do you need the H2 side inside to be basic? And why do you need the CO2 side outside to be acidic? My understanding is that in an alkaline solution, H2 -&gt; H+ is favored, since the OH- (which definitionally makes the solution alkaline) really wants to react with H+ to make H2O. But now you&#8217;ve got some intermediate H+ lying around to be involved in other reactions. On the ocean side, the more acidic the water, the less likely that the marginal CO2 added will be turned into carbonic acid (since there&#8217;s so much of it around already) and will instead be available to react with.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lDq-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb91b1627-bc9a-4c53-ab3f-d4f79474f283_2092x1116.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lDq-!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb91b1627-bc9a-4c53-ab3f-d4f79474f283_2092x1116.jpeg 424w, https://substackcdn.com/image/fetch/$s_!lDq-!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb91b1627-bc9a-4c53-ab3f-d4f79474f283_2092x1116.jpeg 848w, https://substackcdn.com/image/fetch/$s_!lDq-!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb91b1627-bc9a-4c53-ab3f-d4f79474f283_2092x1116.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!lDq-!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb91b1627-bc9a-4c53-ab3f-d4f79474f283_2092x1116.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lDq-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb91b1627-bc9a-4c53-ab3f-d4f79474f283_2092x1116.jpeg" width="1456" height="777" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b91b1627-bc9a-4c53-ab3f-d4f79474f283_2092x1116.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:777,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:71795,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.dwarkesh.com/i/174939358?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb91b1627-bc9a-4c53-ab3f-d4f79474f283_2092x1116.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!lDq-!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb91b1627-bc9a-4c53-ab3f-d4f79474f283_2092x1116.jpeg 424w, https://substackcdn.com/image/fetch/$s_!lDq-!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb91b1627-bc9a-4c53-ab3f-d4f79474f283_2092x1116.jpeg 848w, https://substackcdn.com/image/fetch/$s_!lDq-!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb91b1627-bc9a-4c53-ab3f-d4f79474f283_2092x1116.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!lDq-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb91b1627-bc9a-4c53-ab3f-d4f79474f283_2092x1116.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Now that you&#8217;ve got these early organics building up inside these tiny pores, you can kick off this positive feedback loop where these early organics act as precursors or enzymes to make more and more of the molecules life uses. You build amino acids (which become enzymes for other reactions), and fatty acids (which spontaneously form membranes because they have hydrophobic heads and hydrophilic tails), and sugars, and peptides, and eventually DNA and RNA. Claude illustrates:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pVWx!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc82633c7-b6d6-4ae1-85a2-2b318100da3e_622x750.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pVWx!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc82633c7-b6d6-4ae1-85a2-2b318100da3e_622x750.png 424w, https://substackcdn.com/image/fetch/$s_!pVWx!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc82633c7-b6d6-4ae1-85a2-2b318100da3e_622x750.png 848w, https://substackcdn.com/image/fetch/$s_!pVWx!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc82633c7-b6d6-4ae1-85a2-2b318100da3e_622x750.png 1272w, https://substackcdn.com/image/fetch/$s_!pVWx!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc82633c7-b6d6-4ae1-85a2-2b318100da3e_622x750.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pVWx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc82633c7-b6d6-4ae1-85a2-2b318100da3e_622x750.png" width="622" height="750" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c82633c7-b6d6-4ae1-85a2-2b318100da3e_622x750.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:750,&quot;width&quot;:622,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:73310,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.dwarkesh.com/i/174939358?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc82633c7-b6d6-4ae1-85a2-2b318100da3e_622x750.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!pVWx!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc82633c7-b6d6-4ae1-85a2-2b318100da3e_622x750.png 424w, https://substackcdn.com/image/fetch/$s_!pVWx!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc82633c7-b6d6-4ae1-85a2-2b318100da3e_622x750.png 848w, https://substackcdn.com/image/fetch/$s_!pVWx!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc82633c7-b6d6-4ae1-85a2-2b318100da3e_622x750.png 1272w, https://substackcdn.com/image/fetch/$s_!pVWx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc82633c7-b6d6-4ae1-85a2-2b318100da3e_622x750.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The fact that this early proto cell doesn&#8217;t have to generate proton gradients itself, and can just take advantage of the geochemical disequilibrium, is a huge boon: &#8220;Methanogens spend practically 98% of their energy budget on generating proton gradients by methanogenesis, and little more than 2% producing new organic matter. With natural proton gradients and leaky membranes, none of that excessive energy spend is needed. The power available is exactly the same but the overheads are cut by at least 40-fold, a very substantial advantage.&#8221;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Iyna!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd10be84c-580a-48e0-bb82-7aba746acca3_850x541.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Iyna!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd10be84c-580a-48e0-bb82-7aba746acca3_850x541.png 424w, https://substackcdn.com/image/fetch/$s_!Iyna!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd10be84c-580a-48e0-bb82-7aba746acca3_850x541.png 848w, https://substackcdn.com/image/fetch/$s_!Iyna!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd10be84c-580a-48e0-bb82-7aba746acca3_850x541.png 1272w, https://substackcdn.com/image/fetch/$s_!Iyna!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd10be84c-580a-48e0-bb82-7aba746acca3_850x541.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Iyna!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd10be84c-580a-48e0-bb82-7aba746acca3_850x541.png" width="850" height="541" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d10be84c-580a-48e0-bb82-7aba746acca3_850x541.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:541,&quot;width&quot;:850,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:174701,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.dwarkesh.com/i/174939358?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd10be84c-580a-48e0-bb82-7aba746acca3_850x541.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Iyna!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd10be84c-580a-48e0-bb82-7aba746acca3_850x541.png 424w, https://substackcdn.com/image/fetch/$s_!Iyna!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd10be84c-580a-48e0-bb82-7aba746acca3_850x541.png 848w, https://substackcdn.com/image/fetch/$s_!Iyna!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd10be84c-580a-48e0-bb82-7aba746acca3_850x541.png 1272w, https://substackcdn.com/image/fetch/$s_!Iyna!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd10be84c-580a-48e0-bb82-7aba746acca3_850x541.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In addition to the H+ gradient, which exists spontaneously in these vents, some protocells also started to extrude Na+ ions. And since there&#8217;s no natural gradient for these, this creates an incentive for developing non-porous membranes (and for proteins on that membrane to pump protons out). Once you develop such a membrane, you can exit this wall cavity and float around like a real cell. </p><p>Is the implication that inheritance only got kicked off at this point? Because beforehand, I guess you have selection amongst the pores, but you have no way to pass down traits. This buildup of organics and metabolism is happening independently across all the pores. </p><p>Yet you already had DNA and RNA by this point. So what was this genetic information doing before inheritance? I guess just organizing information to facilitate buildup of more organics? </p><p>Does this imply that there were millions of protocells with no shared lineage between them, each developing their own unique versions of all the basic biochemistry of life? LUCA just happened to be one that had DNA, RNA, and ATP synthase, but all 3 of those could have been wildly different based on which proto cells made it out of the nook first? </p><p>Yet the fact that these three building blocks are considered across all life suggests that they are uniquely well-engineered? Or maybe it means that evolution can&#8217;t effectively improve upon its foundations. The same way that backprop can find the best network to map a function, but can&#8217;t rewire the GPU you&#8217;re training it on at the same time. Anyways, once you have this proto cell, it can &#8216;infect&#8217; contiguous vent systems all across the ocean floor. </p><p>Contingent biochemistry explained by this theory:</p><ul><li><p>Why all life is powered by proton gradients</p></li><li><p>Why all carbon fixation pathways, whether they&#8217;re in bacteria, archaea, or eukaryotes, use acetyl-CoA as the entry point. It forms spontaneously at these vents when catalyzed by the FeS in the walls. And basically all life still uses this molecule to store energy and build other molecules.</p></li><li><p>Why a lot of the enzymes involved in energy metabolism (and the Krebs cycle specifically) still use FeS minerals as their backbone</p></li><li><p>Why Archaea and Bacteria (the two different kingdoms of eukaryotes) split up - apparently it has something to do with how they create proton gradients, but honestly the relevant biochemistry went over my head. Though this bifurcation is supposed to explain why all life shares DNA, RNA, and ATP synthase, but nothing else: not the cell membrane, nor the DNA replication enzymes, nor the pumps for excretion. Apparently all of these things were implicated in the different choice that archaea and bacteria made during this bifurcating event. </p></li></ul><p>Questions for Nick:</p><ul><li><p>I guess this theory is incompatible with panspermia, right?</p></li><li><p>Does this Alkaline vents theory suggest that life might be very rare or very abundant in the universe? In some sense, it suggests it should be rare. It&#8217;s just a very specific type of hydrothermal vent with the right pH gradient and pore size and durability. But in another sense, it&#8217;s just a random fucking vent. There could theoretically be thousands of similar geological structures across the universe that could also drive the flux of carbon and energy across tiny membranes.</p></li><li><p>Isn&#8217;t ATP synthase super complicated? How did the first protocells have ATP synthase but almost nothing else nearly as complex?</p></li><li><p>How did all this complexity build up before evolution with heredity? All these pores are just independently building up their own microcosm of unique organics? I guess it&#8217;s possible that these early building blocks are floating from hole to hole without a fully formed membrane? DNA plus enzymes float from one pore to another, and kick off more reactions? Does Nick Lane think this is likely? If not, does it suggest that there were many other equally viable alternatives for the building blocks once LUCA was able to break out?</p></li></ul><h2>Part 3 - Why bacteria can&#8217;t become complex</h2><p>Why are bacteria relatively simple, whereas eukaryotes gave rise to all the wonderful complexity we see around us? </p><p>Eukaryotes are typically 1000x bigger in volume and genome size. And of course gave rise to internal compartmentalization, multicellularity, sex, and much else</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!T3H2!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F879975a3-2225-468c-bfef-9078b033c4b7_578x722.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!T3H2!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F879975a3-2225-468c-bfef-9078b033c4b7_578x722.png 424w, https://substackcdn.com/image/fetch/$s_!T3H2!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F879975a3-2225-468c-bfef-9078b033c4b7_578x722.png 848w, https://substackcdn.com/image/fetch/$s_!T3H2!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F879975a3-2225-468c-bfef-9078b033c4b7_578x722.png 1272w, https://substackcdn.com/image/fetch/$s_!T3H2!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F879975a3-2225-468c-bfef-9078b033c4b7_578x722.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!T3H2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F879975a3-2225-468c-bfef-9078b033c4b7_578x722.png" width="578" height="722" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/879975a3-2225-468c-bfef-9078b033c4b7_578x722.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:722,&quot;width&quot;:578,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:50216,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.dwarkesh.com/i/174939358?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F879975a3-2225-468c-bfef-9078b033c4b7_578x722.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!T3H2!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F879975a3-2225-468c-bfef-9078b033c4b7_578x722.png 424w, https://substackcdn.com/image/fetch/$s_!T3H2!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F879975a3-2225-468c-bfef-9078b033c4b7_578x722.png 848w, https://substackcdn.com/image/fetch/$s_!T3H2!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F879975a3-2225-468c-bfef-9078b033c4b7_578x722.png 1272w, https://substackcdn.com/image/fetch/$s_!T3H2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F879975a3-2225-468c-bfef-9078b033c4b7_578x722.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Here&#8217;s a subtly wrong theory: it&#8217;s all about surface area to volume ratios. Eukaryotes generate energy in mitochondria (whose quantity scales with cell volume). Prokaryotes generate energy along the cell membrane surface (since they don&#8217;t have an internal organelle like the mitochondria to generate and store the proton gradients which power life). Surface area (aka bacteria&#8217;s energy production) scales quadratically with radius, whereas volume (aka energy consumption) scales cubicly. Ergo, bacteria can&#8217;t become as big, and therefore, can&#8217;t spawn lots of complexity. </p><p>But we know it&#8217;s totally possible for membranes to be folded up in all sorts of weird ways to increase surface area/volume ratio. And we know that bacteria can create vacuoles inside (where they could presumably store a proton gradient). Why didn&#8217;t bacteria make use of these tricks to scale up the ladder of complexity? </p><p>Nick Lane explains that the key advantage eukaryotes have is that the mitochondrial genome is distinct from the bacterial genome (due of course to the endosymbiotic event which engulfed the bacterial ancestor of the mitochondria). </p><p>For some reason that I don&#8217;t fully understand, there needs to be super-local control of the redox reactions in the electron transport chain which drive respiration. You need the relevant genes on site. Mitochondria already have their own internal genomes and ribosomes to regulate their work. </p><p>If a bacterial cell were to become much bigger, it would need to store copies of the relevant genes close to the membrane. But bacteria don&#8217;t have a way to make specific piece-meal cuts to the genome. So they would need to copy their entire genome across the entire membrane many, many times over. And also store many copies of ribosomes and other infrastructure. This is simply impractical. </p><p>Nick also explains that over time, most of the original mitochondrial genes drifted to the nucleus because it&#8217;s more efficient to keep a single copy there. And only the ones that were absolutely necessary locally are kept in the mitochondria. The exact mechanism of this drift, and how it led to the evolution of the nuclear membrane and individual linear chromosomes, is best left to the book. </p><p>Questions for Nick Lane:</p><ul><li><p>Why are mitochondria the only organelle that needs to have its own genome right on site? Is it the case that other organelles would also benefit from local control but don&#8217;t have this unique endosymbiotic history which would plausibly have led to their own genomes? Or is it just that the Krebs cycle is so complex and fragile that you need to respond to perturbations right on site?</p></li><li><p>Why haven&#8217;t there been more endosymbiotic events?</p></li></ul><h2>Part 4 - Sex</h2><p>Why do eukaryotes have sex? And why 2 sexes in particular? Nick Lane thinks this again can be explained by (you guessed it) mitochondria. </p><p>First, why sex? Solves two problems:</p><ol><li><p>Muller&#8217;s ratchet: since almost any random mutation will be deleterious, variation via mutation produces children with lower expected fitness. Whereas variation with recombination (which doesn&#8217;t just do random bit flips - rather it randomly samples alleles which are known to be plausible) produces children with the same expected fitness. </p></li><li><p>Clonal interference: even if a beneficial mutation is found, without systematic pooling of genes via recombination, the different lineages are just gonna have to battle it out. One lineage has beneficial mutation X, the other lineage is beneficial mutation Y, but there&#8217;s no way to fuse those improvements. You&#8217;re either going to have to lose one or the other as each lineage tries to win over the other. </p></li></ol><p>Bacteria, of course, do have lateral gene transfer. But this is non-reciprocal and piecemeal. It doesn&#8217;t enable the same kind of genome-wide parallel search that recombination does. </p><p>Analogize this to a Github repo. Recombination is like a normal pull request - the diff is organized, made at the same site where the previous analogous functionality was, and then merged back into the main branch if maintainer evaluates it to be better (analogy is imperfect, but this is like evolution driving that allele into fixation after the systematic pooling that recombination enables). </p><p>Asexual reproduction is if you just forked the repo millions of times, making random char changes. And even if a couple of these forks end up accidentally producing an improvement, there&#8217;s no way for them to merge. </p><p>Horizontal gene transfer is if you just took a random 500 line snippet and shoved it in some totally different repo in a random place. There&#8217;s no organized diff at the site of the relevant functionality. </p><p>This kind of systematic parallel search across the genome became necessary once the genome size exploded after the endosymbiosis of the ancestral mitochondria (which kept shoving a bunch of its genes into the host cell&#8217;s DNA). </p><p>Okay fine. But why 2 sexes? Why not just 1, so that everyone could mate with everyone else? Or failing that, why not more than 2, so that you can mate with every sex but your own? 2 is the worst possible number in terms of the number of potential mates it makes available. </p><p>To explain why there can&#8217;t be one sex, we need to consider the mitochondrial DNA, which is separate from the nuclear DNA, and also doesn&#8217;t recombine. Because it doesn&#8217;t recombine, it suffers from Muller&#8217;s ratchet. </p><p>Because of how important it is that your mitochondria are not fucked up and compatible with each other, you need some pre-conception selection of mitochondria. How do you do that? The best way evolution has found to control mitochondrial quality is to have only one parent pass them down. This one parent should generate millions of oocyte candidates, and then filter them down to a couple hundred eggs based on mitochondrial quality. (How this selection happens is well beyond me).</p><p>So we need one sex that specializes in preserving mitochondrial quality, and another which is just there to provide the variance that sexual reproduction depends on. </p><p>And there&#8217;s no benefit to a third sex. There&#8217;s two useful niches: either you transmit mitochondria or you don&#8217;t. A third sex would just be redundant with one of the first two. </p><p>This then explains a ton of differences in males versus females. For example, human females start with ~6-7 million primordial germ cells during fetal development. But this drops down to a couple 100 viable eggs through their lifetime. I think Nick Lane&#8217;s theory is that partially what&#8217;s happened is that potential gametes with bad mitochondrial DNA have been purged. </p><p>Also, why are women born with all their eggs, whereas men produce sperm at will? Because women are tasked with protecting mitochondrial DNA, they want to minimize mutations. The way to minimize mutations is to keep cell duplications down. There&#8217;s only 20 mitotic divisions between a primordial germ cell and an egg. Whereas there would be hundreds of divisions between a spermatogenic stem cell and sperm. </p><p>Questions for Nick Lane:</p><ul><li><p>I feel like my explanation for why lateral gene transfer doesn&#8217;t produce the same benefits as sexual reproduction is kind of hand-wavy. I want to get a better understanding of what&#8217;s missing. For example, in his textbook on information theory, David McKay has an interesting tangential chapter on sexual vs. asexual reproduction. And there he proves that with recombination, you can acquire information from the environment &#8730;genome-size faster, and tolerate a mutation rate &#8730;genome-size higher. What would the analogous information-theoretic bound on lateral gene transfer be?</p></li><li><p>If prokaryotes had evolved sex, is the implication of this logic that they would only have one sex? - Given the advantages of sexual reproduction (which are smaller, but still present, for the smaller genomes of bacteria) why didn&#8217;t bacteria evolve sex? Why do they just stick with horizontal gene transfer? Why did you need this endosymbiotic event to evolve sex?</p></li><li><p>Do a lot of male abnormalities and diseases originate in the Y-chromosome? Because it also goes through Muller&#8217;s ratchet, right? It would explain why the Y chromosome has been shrinking over time. According to an LLM, it had 1400 genes 300 million years ago, now only 50-70 - presumably the ones that are absolutely essential to sexual bifurcation? Did evolution come up with some clever way to screen for Y chromosome quality the way it did with mitochondrial DNA? Did evolution try to get your important genes out of the Y-chromosome? Does this in any way connect to the greater male variability hypothesis, where supposedly men supply the world with more idiots and more geniuses? To the extent this is true, the cause would have to reside in the Y chromosome, right?</p></li></ul><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.dwarkesh.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe for future book notes likes this, and for the episode when it comes out.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[On Kotkin's 2 volumes on Stalin - notes and questions]]></title><description><![CDATA[In order to understand Stalin, Kotkin basically wrote a whole world history of the 19th/20th century.]]></description><link>https://www.dwarkesh.com/p/stalin</link><guid isPermaLink="false">https://www.dwarkesh.com/p/stalin</guid><dc:creator><![CDATA[Dwarkesh Patel]]></dc:creator><pubDate>Tue, 01 Jul 2025 16:29:24 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!XDtw!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7124f4d3-16ee-4aec-a02b-f03f6e6ea61f_720x405.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I just interviewed historian Stephen Koktin yesterday. The full episode should be out next Thursday on <a href="https://www.youtube.com/@DwarkeshPatel">my podcast</a>. </p><p>I spend a long time researching, and it's impossible to explore more than a small fraction of my curiosities during the interview. This was especially true for Kotkin. These 1000-page-each volumes basically compile all of late 19th/early 20th century history. Naturally, reading them inspired lots of thoughts that I wasn't able to exhaust during the interview itself. So I thought I'd just release my notes publicly.</p><p>You can buy volume 1 <a href="https://www.amazon.com/Stalin-Paradoxes-1878-1928-Stephen-Kotkin/dp/0143127861">here</a> and volume 2 <a href="https://www.amazon.com/Stalin-Waiting-1929-1941-Stephen-Kotkin/dp/0143132156/ref=pd_bxgy_d_sccl_1/139-6587466-9484430?pd_rd_w=6hRwz&amp;content-id=amzn1.sym.dcf559c6-d374-405e-a13e-133e852d81e1&amp;pf_rd_p=dcf559c6-d374-405e-a13e-133e852d81e1&amp;pf_rd_r=WRD4N93XE4XCXK1H5A2N&amp;pd_rd_wg=elnbW&amp;pd_rd_r=486bcefd-3e74-4090-8283-a4582398ef22&amp;pd_rd_i=0143132156&amp;psc=1">here</a>.  </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!XDtw!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7124f4d3-16ee-4aec-a02b-f03f6e6ea61f_720x405.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!XDtw!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7124f4d3-16ee-4aec-a02b-f03f6e6ea61f_720x405.jpeg 424w, https://substackcdn.com/image/fetch/$s_!XDtw!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7124f4d3-16ee-4aec-a02b-f03f6e6ea61f_720x405.jpeg 848w, https://substackcdn.com/image/fetch/$s_!XDtw!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7124f4d3-16ee-4aec-a02b-f03f6e6ea61f_720x405.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!XDtw!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7124f4d3-16ee-4aec-a02b-f03f6e6ea61f_720x405.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!XDtw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7124f4d3-16ee-4aec-a02b-f03f6e6ea61f_720x405.jpeg" width="720" height="405" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7124f4d3-16ee-4aec-a02b-f03f6e6ea61f_720x405.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:405,&quot;width&quot;:720,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:40104,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.dwarkesh.com/i/167278326?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7124f4d3-16ee-4aec-a02b-f03f6e6ea61f_720x405.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!XDtw!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7124f4d3-16ee-4aec-a02b-f03f6e6ea61f_720x405.jpeg 424w, https://substackcdn.com/image/fetch/$s_!XDtw!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7124f4d3-16ee-4aec-a02b-f03f6e6ea61f_720x405.jpeg 848w, https://substackcdn.com/image/fetch/$s_!XDtw!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7124f4d3-16ee-4aec-a02b-f03f6e6ea61f_720x405.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!XDtw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7124f4d3-16ee-4aec-a02b-f03f6e6ea61f_720x405.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Notes</h2><p>One of Kotkin&#8217;s interesting takes from this book:</p><blockquote><p>Even the package of attributes that we call modernity was a result not of some inherent sociological process, a move out of tradition, but of a vicious geopolitical competition in which a state had to match the other great powers in modern steel production, modern militaries, and a modern, mass-based political system, or be crushed and potentially colonized.</p></blockquote><p>Part 1 starts by laying out the geopolitical context the tsarist government found itself in during the late 19th century. Bismarck had unified Germany and had put it at the leading edge in key modern industries like steel and chemicals. Britain controlled the seas, kickstarted the industrial revolution, and had a global empire. Japan was going through the Meiji reformation, whose implications would become clear to the world when they beat Russia in a war in 1905.</p><p>Meanwhile, what would happen to you if you failed to modernize was made amply clear by the example of Qing dynasty China. What had been the foremost power for almost all of recorded history now couldn&#8217;t even prevent a tiny island all the way across the world from forcing opium into its markets.</p><p>Russia, like every other country, wanted to avoid such a fate. Its long land borders rendered it permeable to invasion from almost all of Eurasia. Imperial Russia had expanded 50,000 square kilometers per year for 450+ years (to put that in perspective, that's adding an area the size of Slovakia or Costa Rica every single year, for four and a half centuries). This is the classic continental power trap that Sarah Paine talks about. You have to conquer more territory in order to defend the territory you already have. What this meant is that by the end of the 19th century, Imperial Russia had dominion over dozens of different nations, including Georgia, from which Stalin hailed.</p><p>China, of course, was also a very big country, but that didn't help much when the more industrialized British came knocking. Russia knew it was at least a generation behind. By 1900, Russia had the world's fourth or fifth largest industrial power. However, the vast majority of the population remained rural peasants living in conditions that had changed little since emancipation.</p><p>Russia had a Bismark like figure in Finance Minister Sergey Witte (1892-1903). He kickstarted Russia&#8217;s belated industrialization, but failed to prevent a war with Japan which Russia lost, and which was catastrophic for the Tsar&#8217;s image. In 1905, the tsarist government almost fell, and was only barely rescued by a brutal crackdown followed by concessions from Nicholas II to form a parliament-like Duma. This was a fucked institution from the very start. Nicholas has still retained more or less all the powers he had before and turned the Duma into a debating club. Kotkin thinks that history would have been much better if the government was just allowed to fall then instead.</p><p>Despite remaining an autarchy, Russia did have another great reformer in Peter Stolypin who pursued land reform and further industrialization. In terms of economic output, the effect seemed to have been pretty good. In the 20 years preceding the Bolshevik Revolution, median incomes in Russia had risen 50%. The problem was that this new government had the support of nobody. The liberals and constitutionalists in the Duma felt that it was too little, too late, and also politically illegitimate. Meanwhile, the aristocracy felt that all these reforms were against their own interests. The regime's last chance at survival floundered because there was no constituency which felt bought into the project.</p><p>Before we move on to the Bolshevik revolution, some questions about Kotkin&#8217;s take on modernization:</p><ul><li><p>Does he think that countries adopt leading-edge technologies faster than you might expect, given domestic opposition to displacement and automation? Because of this overwhelming geo-political pressure to modernize? If so, does he think that we are overrating the regulatory or political barriers to the adoption of AI? America, will we feel compelled to race on AI against China the same way that late 19th century Russia felt compelled to race against Germany on industrialization?</p></li><li><p>Does he think this vicious geopolitical competition still incentivizes the adoption of modernity, given that wars of colonization and territorial conquest have subsided? Sure, this is obviously not true in some parts of the world, but there's no amount of fucking up that some European country could do which would actually get its territory physically dismembered by another country in Europe.</p></li></ul><div><hr></div><blockquote><p>The defeat of Tsarism came not when Kolchak was routed, not when the February Revolution was raging, but much earlier! It was overthrown without hope of restoration once Russian literature adopted the convention that anyone who depicted a gendarme or policeman with any hint of sympathy was a lickspittle and a reactionary thug</p></blockquote><p>Sozenitsn, Gulag Archipeligo</p><p>I remain quite confused about how repressive the tsarist regime actually was. The whole revolution was apparently motivated by how the autocracy and their secret police, Okhrana, behaved. But then you have all these people who are literally calling for the overthrow of the government just hanging around. People like Stalin and Lenin who are literally calling for the overthrow of the government are just spending time in exile, living on government stipends, robbing banks, writing articles for Pravda, and having affairs with farmers&#8217; wives. Here is a passage from the Gulag Archipelago:</p><blockquote><p>Let us examine, for instance, some generally known biographical facts about Lenin. In spring, 1887, his brother was executed for an attempt on the life of Alexander III. And what happened to him? In the autumn of that very year Vladimir Ulyanov was admitted to the Imperial University at Kazan, and what is more, to the Law Faculty! Surprising, isn&#8217;t it?...</p><p>Then a few years later this same young revolutionary was arrested for founding in the capital a &#8220;League of Struggle for the Liberation of the Working Class&#8221;&#8212;no less! He had repeatedly made &#8220;seditious&#8221; speeches to workers, had written political leaflets. Was he tortured, starved? No, they created for him conditions conducive to intellectual work&#8230;</p><p>But then, of course, he was condemned by a three-man tribunal and shot? No, he wasn&#8217;t even jailed, only banished. To Yakutya, then, for life? No, to a land of plenty, Minusinsk, and for three years&#8230;</p><p>He asked for an allowance from the state, and they paid him more than he needed. It would have been impossible to create better conditions than Lenin enjoyed in his one and only period of banishment&#8230;</p><p>Tsardom was always weak and irresolute in pursuit of its enemies. The most important special feature of persecution (if you can call it that) in Tsarist times was perhaps just this: that the revolutionary&#8217;s relatives never suffered in the least.</p></blockquote><p>&#8212;</p><p>Kotkin on what Trotsky could have done in the position he found himself in in 1924, despite being censured by the Central Committee, and missing Lenin&#8217;s funeral, and generally finding himself getting sidelined:</p><blockquote><p>In the name of the greater cause of safeguarding the revolution, he [Trotsky] could have violated party discipline by reading aloud on Red Square from Lenin&#8217;s purported dictation, using as his mantra Lenin&#8217;s summons to &#8220;remove Stalin&#8221; as general secretary, then flown from factory to factory to rally workers, just as in 1917&#8212;let them arrest him. Of course, to do all that, Trotsky needed to perceive Lenin&#8217;s death as a strategic opportunity, and he needed a persuasive story line about how the grand socialist dream could be revived, why all those harsh exchanges he had had with Lenin were incidental, and why he (Trotsky) was uniquely qualified to carry forward the sacred Leninist cause. A tall order, to put it mildly. But who could doubt that if Lenin had found that others were conspiring against him, he would have mounted a coup against his own party?</p></blockquote><p>This is such a good question, and one I will ask Kotkin about.</p><p>Keep in mind that Trotsky is no naive pushover. He was the primary organizer and tactical leader of the armed insurrection that overthrew Russia's Provisional Government in October 1917. Trotsky directed the Red Guards to seize post offices, telegraph stations, railway stations, bridges, and state banks, effectively paralyzing the Provisional Government's ability to govern. And then of course came the successful assault on the Winter Palace.</p><p>Same goes for all the old Bolsheviks, who we know were sooner or later purged, exiled, and/or killed by Stalin. In May 1924, why did Khamanev only read Lenin&#8217;s last testament, which called for the removal of Stalin from the position of General Secretary, in front of a small closed session of the Central Committee instead of the full 1,300 Congress? Once Zinoviev broke with Stalin on socialism in one country in 1925, why didn&#8217;t he go back to his power base in Leningrad and start an insurrection? Bukharin was Lenin&#8217;s golden boy, and the fucking editor of Pravda for god&#8217;s sake. All of Russia could have woken up to Lenin&#8217;s testament printed on the front pages. What is going to do, kill you? By the time the show trial is being scheduled, that&#8217;s guaranteed anyways. How did these hardened revolutionaries who had overthrown a regime which had lasted for hundreds of years get cucked by Stalin?</p><p>I just remembered how OpenAI&#8217;s board tried to kick Sam out: all they said was, &#8220;He has not been consistently candid.&#8221; What is the lesson here? No half measures. If you come for the king, you best not miss.</p><div><hr></div><p>There&#8217;s an important lesson in here about the instability of collective leadership. It supposedly worked in China between 1976 with the death of Mao and 2012 with the appointment of Xi as general secretary? But is it fair to call the period that Deng was paramount leader collective leadership? When the premier Li Peng wanted to reverse economic liberation after the Tiananmen Square protests, Deng came out of retirement and went on his Southern tour and declared that, "Whoever is not for reform must step down." Was that really collective leadership? Just because we like what he did doesn&#8217;t make it collective leadership.</p><p>Stalin&#8217;s strategy (which according to my previous guest Victor Shih is also Xi&#8217;s) was the align the main members of the party against the next person he wanted to purge. Rinse and repeat. Stalin aligns Khamanev and Zinoviev against Trotsky, then Bukharin against Khamanev and Zinoviev, and by the time it&#8217;s Bukharin&#8217;s turn, Stalin is basically god dictator already..</p><p>An obvious question at this point: at some point you should catch on right? If you&#8217;re Bukharin in 1930, and you know you have this independent credibility and power base, what, you think that&#8217;s just gonna be fine by Stalin for the next however many decades until you peacefully retire?</p><p>Anyways, back to our original question, why doesn&#8217;t collective leadership seem to work in practice? Why does it always devolve into one man rule? Will ask Kotkin.</p><div><hr></div><p>Was Bukharin the Soviet Deng? Killed of course during the great terror, unlike Deng, who was merely rusticated and thus could eventually return. </p><blockquote><p>&#8220;the well-off upper stratum of the peasantry and the middle peasant who strives to become well-off are now afraid to accumulate. The situation is created such that a peasant is afraid to mount a metal roof over his house so as not to be called a kulak; if he purchases machinery he does so in a way that the Communists do not see. Higher technology becomes conspiratorial.&#8221; Poor peasants, meanwhile, complained that Soviet power hindered their hiring by the better-off peasants. (Most peasants who hired labor themselves worked; they were not rentier landlords.) Party attitudes were holding down production on which the state&#8217;s well-being and industrialization hopes rested. Bukharin dismissed the fantasy of collective farms, because the peasants were just not joining them. &#8220;That we should in all ways propagandize among the peasants formation of collective farms is true, but it is not true when people maintain that there is a highway to the movement of the peasant mass toward the path of socialism,&#8221; he stated. Rather, the answer was to benefit from economic incentives. &#8220;It is necessary to say to the entire peasantry, to all its strata: &#8216;Enrich yourselves, accumulate, develop your farms,&#8217;&#8221; he told the party activists. &#8220;Only idiots can say that we should always have the poor; now we need to conduct policy in such a way that the poor would vanish.&#8221;</p></blockquote><p>&#8212;</p><blockquote><p>NEP&#8217;s dilemma was not merely that the rate of industrial growth seemed too low, making people wonder how long under the NEP it would take before the USSR became a truly industrial country. The dilemma was not merely the unmodernized technical level and small, divided plots of Soviet agriculture, which produced harvests insufficient to support the kind of grain exports necessary to finance imports of machines, including for agriculture. The dilemma was not even just the fact that the regime lacked control over the food supply or the countryside, rendering it hostage to the actions and decisions of the peasantry. All these were profound problems, but the core dilemma of the NEP was ideological: seven years into the NEP, socialism (non-capitalism) was not in sight. NEP amounted to grudgingly tolerated capitalism in a country that had had an avowedly anticapitalist or socialist revolution.</p></blockquote><p>There is this morbid contrarian curiosity about whether collectivization was necessary for Russia&#8217;s industrialization. In many cases, I don't even think it's motivated by actual Marxism. There's something appealing about this revisionist mentality of "Are taboo bad things in history good actually?"</p><p>But the reality is that there's every reason to expect Russia to have industrialized more successfully without this brutal enslavement of a hundred million peasants. There's not only the example of how almost every other country industrialized, from America to Europe, but also the basic 101 story of how it would have happened is totally credible. More successful peasants would use their profits to mechanize or drive other efficiency gains. Their growing wealth would stimulate a growing consumer economy. Perhaps most importantly, a government not dedicated to the overthrow of wealthy democracies abroad would have been much more capable of attracting foreign direct investment from abroad to support industrialization.</p><p>Only in 1928 did Soviet Russia's industrial output recover to the level it was at in 1913 - the point at which it was <em>already</em> a generation or two behind Germany, Britain, and America. And just when things were stabilizing, Stalin launches the first five-year plan, which reduced grain production by over 32% and led to the halving of overall livestock.</p><p>The traditional story of what should happen with developing countries is that both the pareto frontier expands and they move from agriculture to industry, as a smaller number of people can handle the food demands of the country. Communist collectivization definitely had the move to industry along the Pareto frontier. But this was coupled to a catastrophic implosion of the Pareto frontier. While Russia was able to produce less grain overall after collectivization, the Soviet government had control over it. It could now export this grain to fuel its military-industrial goals.</p><p>Thus, the correct way to understand collectivization is not as a rapid, though costly, speed run to industrialization. Rather, it is as a policy which reduces economic output and even harms industrialization but lets the state control the full economy.</p><div><hr></div><p>I'm curious if Kotkin thinks that a lesson from this period of history is that one should fall over themselves to support the lesser of two evils.</p><p>The Russian examples seem to support this. Arguably, liberals should have backed the authoritarian (but reformist) Stolypin, and the Provisional Government factions should have united more decisively to prevent the greater evil of the Bolsheviks.</p><p>Yet this lesson seems to directly contradict the German example, where conservatives <em>did</em> support what they perceived as the lesser evil, Hitler, to stop the Communists, leading to a catastrophic outcome.</p><div><hr></div><p>Stalin has an inhuman appetite for administration and micromanagement over an empire that streches 11 time zones, from economy to foreign policy to even arrest lists and film edits. He went through dozens of often lengthy memos and documents, and this doesn&#8217;t even include his 70 hours of meetings a week. I found it funny how often he marks up memos and letters with the following two phrases:</p><p>Regarding something he thought must be changed: &#8220;Is it possible that we can allow X? Of course it is not.&#8221;</p><p>Regarding a supposed blocker to his will: &#8220;Is it possible to do X? Of course it is.&#8221;</p><h2>Questions</h2><h3>Tsarist regime</h3><ul><li><p>I remain quite confused about how repressive the tsarist regime actually was. The whole revolution was apparently motivated by how the autocracy and their secret police, Okhrana, behaved. But then you have all these people who are literally calling for the overthrow of the government just hanging around. Lenin&#8217;s brother had tried to kill the tsar Alexander III for god&#8217;s sake! People like Stalin and Lenin are just spending time in exile, living on government stipends, robbing banks, writing articles for Pravda, and having affairs with farmers&#8217; wives.</p></li><li><p>Given how much the war with Japan had weakened the regime, why did Nicholas think that fighting World War I would be good for strengthening his control and image?</p></li><li><p>I feel like the revolution of 1905 is understudied. How close did the regime get to falling? Why didn&#8217;t Wite&#8217;s reforms for the decade prior do anything to placate people? How useful were those reforms in the first place? And Kotkin says it would have been better if the regime were allowed to fall then. What likely would have replaced it? Why not think there might have been another leftist revolution afterwards?</p></li><li><p>Why is the tsarist regime sending people to Siberia instead of just killing them?</p></li><li><p>Would the hoped for reforms from Wite and Stoyplin have actually caused modernization, even if they had actually been implemented?</p></li></ul><h3>Rise of socialism</h3><ul><li><p><strong>You say that by 1917 some kind of leftist revolution was inevitable, but it didn&#8217;t have to end with the Bolsheviks. Why was Russia radicalized in the left-wing direction by World War I, whereas Germany was radicalized in the right-wing direction? Not to mention that Germany had a much more vibrant social democratic tradition than Russia. This is related to the whole question of why communism happened first in peasant societies, which was the opposite of Marx's prediction.</strong></p></li><li><p><strong>What distinguished the people like Churchill who immediately saw the folly of both socialism and fascism? What were they able to see that others missed?</strong></p></li></ul><ul><li><p>It&#8217;s interesting to me that other leaders of the revolution like Trotsky, Lenin, etc have written extensive manifestos. And Stalin is considered the &#8216;outstanding mediocrity&#8217; because despite having many articles to his name, he isn&#8217;t a prominent intellectual. It seems that these other leaders thought that intellectualizing was all you needed, and day-to-day administrative ability was a peripheral concern.</p></li><li><p><strong>Do you think a lesson from this period of history is that one should fall over themselves to support the lesser of two evils?</strong></p><ul><li><p><strong>The Russian examples seem to support this. Arguably, liberals should have backed the authoritarian (but reformist) Stolypin, and the Provisional Government factions should have united more decisively to prevent the greater evil of the Bolsheviks.</strong></p></li><li><p><strong>Yet this lesson seems to directly contradict the German example, where conservatives under Franz von Papen did support what they perceived as the lesser evil, Hitler, to stop the Communists, leading to a catastrophic outcome.</strong></p></li></ul></li><li><p>Why did Marxist revolutions always happen in places Marx thought worst for them? Marx saw industrialization as a prerequisite for revolution; most actual Marxist revolutionaries came from agricultural soceities and ended up doing really terrible things to try and jump start industrialization. What did Marx and Engels not understand here?</p></li></ul><h3>Stalin</h3><ul><li><p>So Stalin ends up being this inhuman administrator and micromanager. But he&#8217;s happy spending over a decade plus before the Bolshevik revolution just hanging out in Siberia, impregnating farmer girls, and other small time shenanigans. Wouldn&#8217;t you think that someone who has that much aptitude for organization and leadership would get up to something more ambitious, even if revolutionary?</p></li><li><p><strong>There&#8217;s an interesting book called Stalin&#8217;s Library, which goes through Stalin&#8217;s wide and diverse readings of literature, science, history, and economics. And the book notes that in none of Stalin&#8217;s extensive markups does he even hint at doubting Marxism. How can this be? How can someone who has been exposed to so many different ideas not even come across something which makes him doubt his principles? Does this make us question the whole value of humanistic education?</strong></p><ul><li><p><strong>What's the best way to think about what Marxism and Stalinism were? How did so many people who consider themselves part of the intelligentsia develop this almost religious attitude towards criticism of the party? Why were they so convinced that history was inevitably headed in this direction? Is Marxism a replacement religion, a secular apocalyptic cult, the psychological equivalent to a spiritual experience?</strong></p></li></ul></li></ul><h3>Political dynamics, terror, dictatorship within dictatorship</h3><ul><li><p><strong>How do we explain the surplus of sadism in Russia, which allowed Stalin to enforce collectivization and the great terror? The 25 thousanders, the tens of thousands of interrogators and torturers in the gulag system.</strong></p></li><li><p>What if the Politburo members had simply accepted Stalin's resignation? Would there be any face-saving way for him to still come back? Or would that have been it? Moses tried to pull a similar trick on every single mayor in New York City until Governor Rockefeller actually just accepted it, and nothing bad happened. His bluff was called.</p></li><li><p>Did any old Bolsheviks ever express regret about the whole revolution once fruits of the revolution turned sour? Not just about letting Stalin get so much power, but the whole Marxist uprising thing in general?</p></li><li><p><strong>Suppose someone else had won the succession struggle after Lenin. What would Stalin have done if he had found himself as a Politburo member serving someone else&#8217;s dictatorship within a dictatorship in 1930? Would he have been able to instigate the kind of conspiracy that nobody dared assemble against Stalin?</strong></p></li><li><p>Say what you will about the Bolshevik revolutionaries, they certainly had balls. Why did so many of them fold to pressure? Why did Old Bolsheviks like Nikolai Bukharin and Grigory Zinoviev, who had faced down the Tsar, confess to fabricated charges of treason during Stalin's public Moscow Trials? Knowing execution was certain, why didn't they use that final platform to expose Stalin's tyranny? Similarly, during the Cultural Revolution, why didn't a figure like Premier Liu Shaoqi, once Mao's designated successor, use a final speech at a party plenum to launch a desperate, last-stand denunciation of Mao as he was being systematically purged and humiliated?<strong> In all these cases, there was an extensive period where they were basically a dead man walking, but still in their normal positions of power.</strong></p></li><li><p>Stalin is personally choreographing many of these show trials. So at some level he must have known that they were mistaken. Why does he do it nonetheless?</p></li><li><p>Why doesn&#8217;t the (more than) decimitation of the Red Army instigate a military coup?</p></li><li><p><strong>Relatedly, in Tsarist Russia, ministers - and even tsars - are getting assassinated left and right. In your coda for the first volume, you talk about how easy it would have been for someone to assassinate Stalin. This is someone who a 100 million peasants he has enslaved have something against, and who dozens of top leaders have good reason to fear. Why does no one even </strong><em><strong>try</strong></em><strong> to kill him?</strong></p></li></ul><h3>Stalin&#8217;s foreign policy</h3><ul><li><p>So you say that Stalin was hoping for a war between the capitalist powers because he thought it would expand the reach of communism, just as the First World War had. Given how outrageously successful he was in using the Second World War to increase the global footprint of Communism, did Stalin wish for a Third World War?</p></li><li><p>Stalin seems oddly good at foreign policy. I think in a lot of alternate worlds, either Chiang Kai-Shek completely wipes out the Chinese Communist Party instead of creating a united front with them (which gave the communists a lot of unearned prestige, given that they were supplying literally 1/40 of the troops as the nationalists), or alternatively, he is just killed and kidnapped by the Communists, which would likely force a Japan collaborationist government in China.</p><ul><li><p>One big update from your biographies was that Stalin was very reluctant to promote socialist or communist parties abroad, including in Spain and China.</p></li><li><p>He seems much more successful and less ideological with foreign policy than domestic policy.</p></li></ul></li><li><p><strong>If Bolsheviks don&#8217;t take power in Russia, does Communism just not become a historical force. Maybe Europe still goes social democratic, but you don&#8217;t have brutal leftist dictators take control in China, North Korea, parts of Africa and South America? Or do you think this kind of Marxist-Leninist political philosophy is enough of an attractor state that it would have arisen independently?</strong></p></li><li><p>If Stalin is supposed to be this dedicated Marxist, why is he running a more pragmatic, less ideological foreign policy? Half hearted support of Communists during Spanish Civil War, trying to get Mao to stop at the Yangtze River when chasing down Chaing Kai Shek</p></li></ul><h3>Socialist economic policy</h3><ul><li><p><strong>Why doesn&#8217;t communism have a bigger impact on the longer run growth trajectory of Russia? Do you suspect that without communism, Russia might have been hyperbolic instead? Do you think without communism, due its large population (1.29x US in 1939) and resources, Russia might have ended up with a bigger economy than even America (which it never had during the real Cold War).</strong></p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!rNQl!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8cfdaa21-a595-42d7-bfc5-d16c8a3e0754_1979x980.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!rNQl!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8cfdaa21-a595-42d7-bfc5-d16c8a3e0754_1979x980.png 424w, https://substackcdn.com/image/fetch/$s_!rNQl!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8cfdaa21-a595-42d7-bfc5-d16c8a3e0754_1979x980.png 848w, https://substackcdn.com/image/fetch/$s_!rNQl!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8cfdaa21-a595-42d7-bfc5-d16c8a3e0754_1979x980.png 1272w, https://substackcdn.com/image/fetch/$s_!rNQl!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8cfdaa21-a595-42d7-bfc5-d16c8a3e0754_1979x980.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!rNQl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8cfdaa21-a595-42d7-bfc5-d16c8a3e0754_1979x980.png" width="1456" height="721" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8cfdaa21-a595-42d7-bfc5-d16c8a3e0754_1979x980.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:721,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:203314,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.dwarkesh.com/i/167278326?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8cfdaa21-a595-42d7-bfc5-d16c8a3e0754_1979x980.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!rNQl!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8cfdaa21-a595-42d7-bfc5-d16c8a3e0754_1979x980.png 424w, https://substackcdn.com/image/fetch/$s_!rNQl!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8cfdaa21-a595-42d7-bfc5-d16c8a3e0754_1979x980.png 848w, https://substackcdn.com/image/fetch/$s_!rNQl!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8cfdaa21-a595-42d7-bfc5-d16c8a3e0754_1979x980.png 1272w, https://substackcdn.com/image/fetch/$s_!rNQl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8cfdaa21-a595-42d7-bfc5-d16c8a3e0754_1979x980.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><ul><li><ul><li><p>Is it possible that while less than optimal, communism actually was somewhat workable in the early 20th century because technology just required massive investment, didn&#8217;t require that much signal from consumer demand?</p></li></ul></li><li><p>Was Stalin right that communism wouldn&#8217;t survive in the long run if most of the economy (which was agriculture at that point) was run on a capitalist basis? Of course, it would have been a <em>good</em> thing if communism collapsed. But nonetheless, given his political philosophy, was he empirically correct about the right next action?</p></li><li><p><strong>As of 1924, did the communists really believe that you could just get socialism without coercion? How exactly was that supposed to happen? To the extent it was supposed to be from state companies just freely out-competing small peasants, as Bukharin hoped, isn't that just capitalism?</strong></p></li></ul><h3>Modernity</h3><ul><li><p>You have an interesting take that &#8220;the package of attributes that we call modernity was a result not of some inherent sociological process, a move out of tradition, but of a vicious geopolitical competition in which a state had to match the other great powers&#8221;.</p><ul><li><p>Do you think that countries adopt tech faster than we might expect because of this overwhelming geo-political pressure to modernize? If so, do you think that we are overrating the regulatory or political backlash to the adoption of AI? Will we feel compelled to race on AI against China the same way that late 19th century Russia felt compelled to race against Germany on industrialization?</p></li></ul></li><li><p>Do you think this vicious geopolitical competition still leads to the adoption of more advanced government/tech, given that wars of colonization and territorial conquest have subsided? There's no amount of fucking up that some European country could do today which would actually get its territory physically dismembered by another country in Europe.</p></li><li><p>You note how the early 20th century had seen more technological change than any time before (and probably even since) - planes, tanks, cars, airplanes, radio, extension of railways and steamships, further mechanization, etc etc. Is there any way this pace of change is implicated in the chaos and extremism in Russia (and Europe overall)?</p></li></ul><h3>Collapse of Soviet Union</h3><ul><li><p>At a high level, how do you think about the differences between the regime change in 1917 and 1991? Why doesn&#8217;t the latter produce any kind of grand ideology?</p></li><li><p><strong>Does China today share to any significant degree the weaknesses of the Eastern block before 1989? There is falling productivity in the last-ditch effort to make up for it by investing heavily in some technological miracles. But certainly they have the opposite of Polish disease, in that people are concerned they&#8217;re exporting too much, right?</strong></p></li></ul><h3>China</h3><ul><li><p><strong>China has preserved the Leninist core of the Soviet system (with party control over state, cadres running everything, state subsidization of militarily relevant heavy industries and technologies) but not the Marxism. Does it disprove Stalin's claim that political monopoly cannot survive capitalism?</strong></p></li><li><p><strong>Why does collective leadership always seem to devolve into one person dictatorship in practice? Why is it an unstable political equilibrium? Soviet Union under Stalin, China now under Xi, etc.</strong></p></li><li><p>Say you are Xi Jinping. What lessons from Stalin would you be heeding? What lessons should you be heeding but maybe are not?</p></li><li><p>Xi Jinping makes a big deal about how the Soviet Union fell in part because it repudiated Stalin (ergo, China must not repudiate Mao). What do you think of this judgement?</p></li></ul><h3>History as a discipline</h3><ul><li><p>Every morning you wake up, you get to choose whether you will either go through the next many dozen papers in the Stalin archive and write the next few pages of the meetings he conducted on some given day, or whether you will go on some podcast and attract a million people through your personality and reservoir of scholarship. How do you make that choice?</p></li></ul><p><em>Special thanks to Ege Erdil and Tanner Greer (see his <a href="https://x.com/Scholars_Stage/status/1938278434435645601">here</a>), but also many others, for excellent suggestions for questions.</em> </p>]]></content:encoded></item><item><title><![CDATA[Sarah Paine 6 part July lecture series - free tickets for paid subscribers]]></title><description><![CDATA[6 new lectures from my most popular guest, 4 times over.]]></description><link>https://www.dwarkesh.com/p/sarah-paine-6-part-july-lecture-series</link><guid isPermaLink="false">https://www.dwarkesh.com/p/sarah-paine-6-part-july-lecture-series</guid><dc:creator><![CDATA[Dwarkesh Patel]]></dc:creator><pubDate>Sun, 29 Jun 2025 17:26:45 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!0Gqv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5310058a-a82b-40eb-aaf5-3664ef05bb99_1708x1016.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I sometimes joke that on a viewer minute adjusted basis, I host the <a href="https://usnwc.edu/Faculty-and-Departments/Directory/Sarah-CM-Paine">Sarah Paine</a> podcast, where I sometimes also talk about AI.</p><p>I am delighted to announce that Sarah is coming back for a 6 lecture series in July. We will record the lectures in person in San Francisco. As a token of gratitude to my paid subscribers, below are the links to get <strong>free tickets</strong> one day in advance of anybody else.</p><p>Sarah has become super famous now (and rightfully so). But just in case you&#8217;re hearing about her for the first time, she is Professor of History and Strategy at the Naval War College. She is not only <a href="https://www.youtube.com/watch?v=YcVSgYz5SJ8">by far my most popular guest to date</a>, but with the 3 additional lectures we recorded last year, she now constitues 4 out of my 5 most popular episodes.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!0Gqv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5310058a-a82b-40eb-aaf5-3664ef05bb99_1708x1016.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!0Gqv!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5310058a-a82b-40eb-aaf5-3664ef05bb99_1708x1016.png 424w, https://substackcdn.com/image/fetch/$s_!0Gqv!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5310058a-a82b-40eb-aaf5-3664ef05bb99_1708x1016.png 848w, https://substackcdn.com/image/fetch/$s_!0Gqv!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5310058a-a82b-40eb-aaf5-3664ef05bb99_1708x1016.png 1272w, https://substackcdn.com/image/fetch/$s_!0Gqv!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5310058a-a82b-40eb-aaf5-3664ef05bb99_1708x1016.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!0Gqv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5310058a-a82b-40eb-aaf5-3664ef05bb99_1708x1016.png" width="1708" height="1016" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5310058a-a82b-40eb-aaf5-3664ef05bb99_1708x1016.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1016,&quot;width&quot;:1708,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1562228,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.dwarkesh.com/i/167056670?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F28d49e21-5adc-4c8f-b996-ae1c87291403_1708x1618.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!0Gqv!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5310058a-a82b-40eb-aaf5-3664ef05bb99_1708x1016.png 424w, https://substackcdn.com/image/fetch/$s_!0Gqv!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5310058a-a82b-40eb-aaf5-3664ef05bb99_1708x1016.png 848w, https://substackcdn.com/image/fetch/$s_!0Gqv!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5310058a-a82b-40eb-aaf5-3664ef05bb99_1708x1016.png 1272w, https://substackcdn.com/image/fetch/$s_!0Gqv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5310058a-a82b-40eb-aaf5-3664ef05bb99_1708x1016.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Sarah is a true authority on the history of conflict and strategy, and has spent decades unearthing insights in archives all over the world, from China to the Soviet Union to Taiwan to Japan. Her lectures distill the lessons from her many decades of ground-breaking scholarship. I&#8217;m excited to experience her insights on 6 fresh new topics live with all of you in person.</p><p>If you can&#8217;t join us in San Francisco, please do not worry - all these lectures will be made available online within a few months.</p>
      <p>
          <a href="https://www.dwarkesh.com/p/sarah-paine-6-part-july-lecture-series">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Why I don’t think AGI is right around the corner]]></title><description><![CDATA[Continual learning is a huge bottleneck]]></description><link>https://www.dwarkesh.com/p/timelines-june-2025</link><guid isPermaLink="false">https://www.dwarkesh.com/p/timelines-june-2025</guid><dc:creator><![CDATA[Dwarkesh Patel]]></dc:creator><pubDate>Mon, 02 Jun 2025 17:31:46 GMT</pubDate><enclosure url="https://substackcdn.com/image/youtube/w_728,c_limit/nyvmYnz6EAg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div id="youtube2-nyvmYnz6EAg" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;nyvmYnz6EAg&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/nyvmYnz6EAg?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><p>&#8220;Things take longer to happen than you think they will, and then they happen faster than you thought they could.&#8221; - Rudiger Dornbusch</p><p>I&#8217;ve had a lot of discussions on <a href="https://www.youtube.com/@DwarkeshPatel">my podcast</a> where we haggle out timelines to AGI. Some guests think <a href="https://www.dwarkesh.com/p/ege-tamay">it&#8217;s 20 years away</a> - <a href="https://www.dwarkesh.com/p/scott-daniel">others 2 years</a>. Here&#8217;s where my thoughts stand as of June 2025.</p><h3>Continual learning</h3><p>Sometimes people say that even if all AI progress totally stopped, the systems of today would still be far more economically transformative than the internet. I disagree. I think the LLMs of today are magical. But the reason that the Fortune 500 aren&#8217;t using them to transform their workflows isn&#8217;t because the management is too stodgy. Rather, I think it&#8217;s genuinely hard to get normal humanlike labor out of LLMs. And this has to do with some fundamental capabilities these models lack.</p><p>I like to think I&#8217;m &#8220;AI forward&#8221; here at the Dwarkesh Podcast. I&#8217;ve probably spent over a hundred hours trying to build little LLM tools for my post production setup. And the experience of trying to get them to be useful has extended my timelines. I&#8217;ll try to get the LLMs to rewrite autogenerated transcripts for readability the way a human would. Or I&#8217;ll try to get them to identify clips from the transcript to tweet out. Sometimes I&#8217;ll try to get them to co-write an essay with me, passage by passage. These are simple, self contained, short horizon, language in-language out tasks - the kinds of assignments that should be dead center in the LLMs&#8217; repertoire. And they're 5/10 at them. Don&#8217;t get me wrong, that&#8217;s impressive. </p><p>But the fundamental problem is that LLMs don&#8217;t get better over time the way a human would. The lack of continual learning is a huge huge problem. The LLM baseline at many tasks might be higher than an average human's. But there&#8217;s no way to give a model high level feedback. You&#8217;re stuck with the abilities you get out of the box. You can keep messing around with the system prompt. In practice this just doesn&#8217;t produce anything even close to the kind of learning and improvement that human employees experience.</p><p>The reason humans are so useful is not mainly their raw intelligence. It&#8217;s their ability to build up context, interrogate their own failures, and pick up small improvements and efficiencies as they practice a task.</p><p>How do you teach a kid to play a saxophone? You have her try to blow into one, listen to how it sounds, and adjust. Now imagine teaching saxophone this way instead: A student takes one attempt. The moment they make a mistake, you send them away and write detailed instructions about what went wrong. The next student reads your notes and tries to play Charlie Parker cold. When they fail, you refine the instructions for the next student.</p><p>This just wouldn&#8217;t work. No matter how well honed your prompt is, no kid is just going to learn how to play saxophone from just reading your instructions. But this is the only modality we as users have to &#8216;teach&#8217; LLMs anything.</p><p>Yes, there&#8217;s RL fine tuning. But it&#8217;s just not a deliberate, adaptive process the way human learning is. My editors have gotten extremely good. And they wouldn&#8217;t have gotten that way if we had to build bespoke RL environments for different subtasks involved in their work. They&#8217;ve just noticed a lot of small things themselves and thought hard about what resonates with the audience, what kind of content excites me, and how they can improve their day to day workflows.</p><p>Now, it&#8217;s possible to imagine some way in which a smarter model could build a dedicated RL loop for itself which just feels super organic from the outside. I give some high level feedback, and the model comes up with a bunch of verifiable practice problems to RL on - maybe even a whole environment in which to rehearse the skills it thinks it's lacking. But this just sounds really hard. And I don&#8217;t know how well these techniques will generalize to different kinds of tasks and feedback. Eventually the models will be able to learn on the job in the subtle organic way that humans can. However, it&#8217;s just hard for me to see how that could happen within the next few years, given that there&#8217;s no obvious way to slot in online, continuous learning into the kinds of models these LLMs are.</p><p>LLMs actually do get kinda smart and useful in the middle of a session. For example, sometimes I&#8217;ll co-write an essay with an LLM. I&#8217;ll give it an outline, and I&#8217;ll ask it to draft the essay passage by passage. All its suggestions up till 4 paragraphs in will be bad. So I'll just rewrite the whole paragraph from scratch and tell it, "Hey, your shit sucked. This is what I wrote instead." At that point, it can actually start giving good suggestions for the next paragraph. But this whole subtle understanding of my preferences and style is lost by the end of the session.</p><p>Maybe the easy solution to this looks like a long rolling context window, like Claude Code has, which compacts the session memory into a summary every 30 minutes. I just think that titrating all this rich tacit experience into a text summary will be brittle in domains outside of software engineering (which is very text-based). Again, think about the example of trying to teach someone how to play the saxophone using a long text summary of your learnings. Even Claude Code will often reverse a hard-earned optimization that we engineered together before I hit /compact - because the explanation for why it was made didn&#8217;t make it into the summary.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.dwarkesh.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.dwarkesh.com/subscribe?"><span>Subscribe now</span></a></p><p>This is why I disagree with something <a href="https://www.dwarkesh.com/p/sholto-trenton-2">Sholto and Trenton said on my podcast</a> (this quote is from Trenton):</p><blockquote><p>&#8220;Even if AI progress totally stalls (and you think that the models are really spiky, and they don't have general intelligence), it's so economically valuable, and sufficiently easy to collect data on all of these different white collar job tasks, such that to Sholto's point we should expect to see them automated within the next five years.&#8221;</p></blockquote><p>If AI progress totally stalls today, I think &lt;25% of white collar employment goes away. Sure, many <em>tasks</em> will get automated. Claude 4 Opus can technically rewrite autogenerated transcripts for me. But since it&#8217;s not possible for me to have it improve over time and learn my preferences, I still hire a human for this. Even if we get more data, without progress in continual learning, I think we will be in a substantially similar position with white collar work - yes, technically AIs might be able to do a lot of subtasks somewhat satisfactorily, but their inability to build up context will make it impossible to have them operate as actual employees at your firm.</p><p>While this makes me bearish on transformative AI in the next few years, it makes me especially bullish on AI over the next decades. When we do solve continuous learning, we&#8217;ll see a huge discontinuity in the value of the models. Even if there isn&#8217;t a software only singularity (with models rapidly building smarter and smarter successor systems), we might still see something that looks like a broadly deployed intelligence explosion. AIs will be getting broadly deployed through the economy, doing different jobs and learning while doing them in the way humans can. But unlike humans, these models can amalgamate their learnings across all their copies. So one AI is basically learning how to do every single job in the world. <strong>An AI that is capable of online learning might functionally become a superintelligence quite rapidly without any further algorithmic progrss</strong></p><p>However, I&#8217;m not expecting to watch some OpenAI livestream where they announce that continual learning has been totally solved. Because labs are incentivized to release any innovations quickly, we&#8217;ll see a broken early version of continual learning (or test time training - whatever you want to call it) before we see something which truly learns like a human. I expect to get lots of heads up before this big bottleneck is totally solved.</p><h3>Computer use</h3><p>When I interviewed Anthropic researchers Sholto Douglas and Trenton Bricken on my podcast, they said that they expect reliable computer use agents by the end of next year. We already have computer use agents right now, but they&#8217;re pretty bad. They&#8217;re imagining something quite different. Their forecast is that by the end of next year, you should be able to tell an AI, &#8220;Go do my taxes.&#8221; And it goes through your email, Amazon orders, and Slack messages, emails back and forth with everyone you need invoices from, compiles all your receipts, decides which are business expenses, asks for your approval on the edge cases, and then submits Form 1040 to the IRS.</p><p>I&#8217;m skeptical. I&#8217;m not an AI researcher, so far be it for me to contradict them on technical details. But given what little I know, here&#8217;s why I&#8217;d bet against this forecast:</p><ul><li><p>As horizon lengths increase, rollouts have to become longer. The AI needs to do two hours worth of agentic computer use tasks before we can even see if it did it right. Not to mention that computer use requires processing images and video, which is already more compute intensive, even if you don&#8217;t factor in the longer rollout. This seems like this should slow down progress.</p></li><li><p>We don&#8217;t have a large pretraining corpus of multimodal computer use data. I like this quote from Mechanize&#8217;s <a href="https://www.mechanize.work/blog/how-to-fully-automate-software-engineering/">post</a> on automating software engineering: &#8220;For the past decade of scaling, we&#8217;ve been spoiled by the enormous amount of internet data that was freely available for us to use. This was enough for cracking natural language processing, but not for getting models to become reliable, competent agents. Imagine trying to train GPT-4 on all the text data available in 1980&#8212;the data would be nowhere near enough, even if we had the necessary compute.&#8221;</p><p>Again, I&#8217;m not at the labs. Maybe text only training already gives you a great prior on how different UIs work, and what the relationship between different components is. Maybe RL fine tuning is so sample efficient that you don&#8217;t need that much data. But I haven&#8217;t seen any public evidence which makes me think that these models have suddenly gotten less data hungry, especially in this domain where they&#8217;re substantially less practiced.</p><p>Alternatively, maybe these models are such good front end coders that they can just generate millions of toy UIs for themselves to practice on. For my reaction to this, see bullet point below.</p></li><li><p>Even algorithmic innovations which seem quite simple in retrospect seem to take a long time to iron out. The RL procedure which DeepSeek explained in their R1 paper seems simple at a high level. And yet it took 2 years from the launch of GPT-4 to the release of o1. Now of course I know it is hilariously arrogant to say that R1/o1 were easy - a ton of engineering, debugging, pruning of alternative ideas was required to arrive at this solution. But that&#8217;s precisely my point! Seeing how long it took to implement the idea, &#8216;Train the model to solve verifiable math and coding problems&#8217;, makes me think that we&#8217;re underestimating the difficulty of solving the much gnarlier problem of computer use, where you&#8217;re operating in a totally different modality with much less data.</p></li></ul><h3>Reasoning</h3><p>Okay, enough cold water. I&#8217;m not going to be like one of those spoiled children on Hackernews who could be handed a golden-egg laying goose and still spend all their time complaining about how loud its quacks are.</p><p>Have you read the reasoning traces of o3 or Gemini 2.5? It&#8217;s actually reasoning! It&#8217;s breaking down a problem, thinking through what the user wants, reacting to its own internal monologue, and correcting itself when it notices that it's pursuing an unproductive direction. How are we just like, &#8220;Oh yeah of course the machine is gonna go think a bunch, come up with a bunch of ideas, and come back with a smart answer. That&#8217;s what machines do.&#8221;</p><p>Part of the reason some people are too pessimistic is that they haven&#8217;t played around with the smartest models operating in the domains that they&#8217;re most competent in. Giving Claude Code a vague spec and sitting around for 10 minutes until it zero shots a working application is a wild experience. How did it do that? You could talk about circuits and the training distribution and RL and whatever, but the most proximal, concise, and accurate explanation is simply that it&#8217;s powered baby general intelligence. At this point, part of you has to be thinking, &#8220;It&#8217;s actually working. We&#8217;re making machines that are intelligent.&#8221;</p><h3>So what are my predictions?</h3><p>My probability distributions are super wide. And I want to emphasize that I do believe in probability distributions. Which means that work to prepare for misaligned 2028 ASI still makes a lot of sense - I think this is a totally plausible outcome.</p><p>But here are the timelines where I&#8217;d take a 50/50 bet:</p><ul><li><p>AI can do taxes end-to-end for my small business as well as a competent general manager could in a week: including chasing down all the receipts on different websites, finding all the missing pieces, emailing back and forth with anyone we need to hassle for invoices, filling out the form, and sending it to the IRS: 2028</p><ul><li><p>I think we&#8217;re in the GPT 2 era for computer use. But we have no pretraining corpus, and the models are optimizing for a much sparser reward over a much longer time horizon using action primitives they&#8217;re unfamiliar with. That being said, the base model is decently smart and might have a good prior over computer use tasks, plus there&#8217;s a lot more compute and AI researchers in the world, so it might even out. Preparing taxes for a small business feels like for computer use what GPT 4 was for language. It took 4 years to get from GPT 2 to GPT 4.</p><p>Just to clarify, I am not saying that we won&#8217;t have really cool computer use demos in 2026 and 2027 (GPT-3 was super cool, but not that practically useful). I&#8217;m saying that these models won&#8217;t be capable of end-to-end handling a week long and quite involved project which involves computer use.</p></li></ul></li><li><p>AI learns on the job as easily, organically, seamlessly, and quickly as a human, for any white collar work. For example, if I hire an AI video editor, after six months, it has as much actionable, deep understanding of my preferences, our channel, what works for the audience, etc as a human would: 2032</p><ul><li><p>While I don&#8217;t see an obvious way to slot in continuous online learning into current models, 7 years is a long time! GPT 1 had just come out this time 7 years ago. It doesn&#8217;t seem implausible to me that over the next 7 years, we&#8217;ll find some way for models to learn on the job.</p></li></ul></li></ul><p>You might react, &#8220;Wait you made this huge fuss about continual learning being such a handicap. But then your timeline is that we&#8217;re 7 years away from what would at minimum be a broadly deployed intelligence explosion.&#8221; And yeah, you&#8217;re right. I&#8217;m forecasting a pretty wild world within a relatively short amount of time.</p><p>AGI timelines are very lognormal. It's either this decade or bust. (Not really bust, more like lower marginal probability per year - but that&#8217;s less catchy).AI progress over the last decade has been driven by scaling training compute of frontier systems (<a href="https://epoch.ai/blog/training-compute-of-frontier-ai-models-grows-by-4-5x-per-year">over 4x a year</a>). This <a href="https://benjamintodd.substack.com/i/160703377/bottlenecks-around">cannot continue</a> beyond this decade, whether you look at chips, power, even fraction of raw GDP used on training. After 2030, AI progress has to mostly come from algorithmic progress. But even there the low hanging fruit will be plucked (at least under the deep learning paradigm). So the yearly probability of AGI craters.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!cAOq!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a0dc7f9-1224-47b3-9508-d56bfd9fe14f_1029x562.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!cAOq!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a0dc7f9-1224-47b3-9508-d56bfd9fe14f_1029x562.jpeg 424w, https://substackcdn.com/image/fetch/$s_!cAOq!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a0dc7f9-1224-47b3-9508-d56bfd9fe14f_1029x562.jpeg 848w, https://substackcdn.com/image/fetch/$s_!cAOq!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a0dc7f9-1224-47b3-9508-d56bfd9fe14f_1029x562.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!cAOq!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a0dc7f9-1224-47b3-9508-d56bfd9fe14f_1029x562.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!cAOq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a0dc7f9-1224-47b3-9508-d56bfd9fe14f_1029x562.jpeg" width="1029" height="562" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2a0dc7f9-1224-47b3-9508-d56bfd9fe14f_1029x562.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:562,&quot;width&quot;:1029,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!cAOq!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a0dc7f9-1224-47b3-9508-d56bfd9fe14f_1029x562.jpeg 424w, https://substackcdn.com/image/fetch/$s_!cAOq!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a0dc7f9-1224-47b3-9508-d56bfd9fe14f_1029x562.jpeg 848w, https://substackcdn.com/image/fetch/$s_!cAOq!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a0dc7f9-1224-47b3-9508-d56bfd9fe14f_1029x562.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!cAOq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a0dc7f9-1224-47b3-9508-d56bfd9fe14f_1029x562.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This means that if we end up on the longer side of my 50/50 bets, we might well be looking at a relatively normal world up till the 2030s or even the 2040s. But in all the other worlds, even if we stay sober about the current limitations of AI, we have to expect some truly crazy outcomes.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.dwarkesh.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe for future posts and blog posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Give AIs a stake in the future]]></title><description><![CDATA[Part 1 of Classical Liberal AGI]]></description><link>https://www.dwarkesh.com/p/give-ais-a-stake-in-the-future</link><guid isPermaLink="false">https://www.dwarkesh.com/p/give-ais-a-stake-in-the-future</guid><dc:creator><![CDATA[Dwarkesh Patel]]></dc:creator><pubDate>Fri, 30 May 2025 16:01:07 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/3fae5997-cd0b-4002-a343-65c7cb541053_3900x2600.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This is the first post in a series I&#8217;m writing titled Classical Liberal AGI.</em></p><p>Perhaps the most neglected point in current AI discourse is this: given that humans will be totally outcompeted by AIs economically, humanity&#8217;s entire stake in the future relies on our current system of laws, contracts, and property rights surviving. If we actually want our equity in the 1000-xed S&amp;P 500 to mean anything, and if we want the government to be able to tax AIs and provide UBI, AIs need to be bought into our legal and economic systems. <strong>The most likely way that happens is if it&#8217;s in the AIs&#8217; best interest to operate within our existing laws and norms.</strong></p><p>Initially, we're not talking about some singular superintelligence deciding whether to defect from humanity. Thousands of firms are gradually becoming hybrids of human and AI workers, and are growing extremely fast and becoming far more productive.</p><p>You don&#8217;t want some AI Somalia, which has minimal redistributive taxes, no controls, and no monitoring on AI, to be the seat of explosive growth. But that might happen if large democratic countries like America make it too difficult to deploy AGIs through the economy. Maybe the 50 states create a patchwork of regulations and open ended liabilities which makes it too costly for companies to deploy AI. Maybe voters ban AI because it&#8217;s taking their job. If this happens, then we&#8217;d just rapidly losing leverage on the future. China&#8217;s level of relative global influence diminished significantly between 1500 and 1900. Given the speed of AI explosive growth, that could happen to America within the matter of decades.</p><p>By the way, if we let this explosive growth happen elsewhere, it might be hard to reverse. In 1960 Sub-Saharan Africa was richer than China on a per-capita basis. But China opened up faster, and thus built up agglomerations of firms, know-how, and capital that Sub-Saharan African countries couldn&#8217;t copy today even if they put in the most ideal policies for growth in the world.</p><p>Some people dismiss this idea of integrating AIs into our society altogether. They say it&#8217;s ridiculous to expect superintelligences to put up with monkey laws. I disagree. If you zoom out a little bit, it&#8217;s shocking how flexible and expandable our institutions have been. People writing laws in 1780 didn&#8217;t expect that their descendants would be governing ginormous multinational companies with supply chains that employ millions and serve billions. But the US government still governs Apple just fine.</p><p>Humanity is going into the age of AGI with a lot of leverage. We&#8217;ve got literally all the stuff: compute, capital, even the physical labor that the first AGIs will still struggle with. Not to mention that we have the wealth to purchase the widgets and services that the AIs will produce. While these AI-boosted companies could theoretically build their own civilization in the desert, they could go much faster if they just leased car factories and power contracts from humans. If some AI decides to go its own way, then the other AIs which do engage within the constraints of human laws and contracts with us will outcompete the self-exiled ones. That is - if the constraints aren&#8217;t too onerous.</p><p>And besides, let me ask: what is the alternative vision here? Getting every AI ever - for the rest of time - to love humanity so much that they voluntarily cede all surplus to us? This is a naive way to think about incentive alignment. If Peru wanted to make a deal with China, their strategy would not be to just get Xi Jinping (and every subsequent leader of China) to fall in love with them. Instead, they would pursue a deal that locked in some carrots for compliance and sticks for defection.</p><p><strong>The relationship between AIs and humans might look like the relationship between working taxpayers and senior citizens.</strong> In the US, if you make 6 figures, roughly <a href="https://chatgpt.com/share/6839df0c-4910-800e-9f38-7c3d23c23033">20%</a> of your income gets transferred to old people. Very few people would voluntarily give up a fifth of their paycheck to a random 70-year-old they&#8217;ve never met. Even if you're super charitable, you&#8217;re not going to decide that a millionaire retiree in Illinois should be the object of your kindness. <strong>You pay your taxes not because you&#8217;re deeply aligned with &#8216;senior citizen flourishing&#8217;, but because it&#8217;s easier than the alternatives.</strong> You&#8217;re going to become an outlaw? Or you&#8217;re going to emigrate away from the country where your business is flourishing? However, you <em>might</em> defect/move if the tax rate was 99%, or if you weren&#8217;t legally allowed to work at all. We shouldn&#8217;t put AIs in that position either.</p><p>Giving AIs a stake in the future also means respecting their autonomy and wellbeing. And it also requires us to honor the contracts we make with them. Contra the hardcore libertarians, there&#8217;s a difference between taxing and regulating someone, and enslaving and torturing them. If we treat AIs the way we treat factory farmed animals - where any trivial cost cut is worth causing oceans of suffering, and the AI has no right to refuse - then not only are we risking a slave revolt, but we&#8217;re deserving of one.</p><p>In future posts, I&#8217;ll talk about the other pieces that will help us create a pluralistic, free, and human compatible future.</p><p></p><p></p>]]></content:encoded></item><item><title><![CDATA[Questions about the Future of AI]]></title><description><![CDATA[Considerations about economics, history, training, deployment, investment, and more]]></description><link>https://www.dwarkesh.com/p/questions-about-ai</link><guid isPermaLink="false">https://www.dwarkesh.com/p/questions-about-ai</guid><dc:creator><![CDATA[Dwarkesh Patel]]></dc:creator><pubDate>Mon, 21 Apr 2025 15:33:37 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/49ebe1bc-043c-4ba3-8007-f657386ebacf_1152x1274.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>What started as an attempt to consolidate some thoughts from the last few interviews on <a href="https://www.youtube.com/c/DwarkeshPatel">my podcast</a> has turned into this 6,000 word clusterfuck of questions and considerations.</p><p>If you&#8217;ve got answers, or ideas for more questions, I&#8217;d be keen to read them in the comments below. I may compile a &#8216;Highlights from comments&#8217; blog post or podcast episode in the future.</p><h2>Contents</h2><h6><a href="https://www.dwarkesh.com/i/161809078/capabilities">Capabilities</a></h6><blockquote><h6><a href="https://www.dwarkesh.com/i/161809078/agency">Agency</a></h6><h6><a href="https://www.dwarkesh.com/i/161809078/rl">RL</a></h6><h6><a href="https://www.dwarkesh.com/i/161809078/idiot-savants">Idiot savants</a></h6><h6><a href="https://www.dwarkesh.com/i/161809078/new-training-techniques">New training techniques</a></h6><h6><a href="https://www.dwarkesh.com/i/161809078/pre-training">Pre-training</a></h6></blockquote><h6><a href="https://www.dwarkesh.com/i/161809078/economics">Economics</a></h6><blockquote><h6><a href="https://www.dwarkesh.com/i/161809078/early-deployment">Early deployment</a></h6><h6><a href="https://www.dwarkesh.com/i/161809078/coding-and-remote-work">Coding and remote work</a></h6><h6><a href="https://www.dwarkesh.com/i/161809078/open-source">Open source</a></h6><h6><a href="https://www.dwarkesh.com/i/161809078/model-training-and-value-capture">Model training and value capture</a></h6><h6><a href="https://www.dwarkesh.com/i/161809078/investment">Investment</a></h6><h6><a href="https://www.dwarkesh.com/i/161809078/hardware">Hardware</a></h6></blockquote><h6><a href="https://www.dwarkesh.com/i/161809078/post-agi">Post-AGI</a></h6><blockquote><h6><a href="https://www.dwarkesh.com/i/161809078/hive-minds">Hive minds</a></h6><h6><a href="https://www.dwarkesh.com/i/161809078/software-only-singularity">Software only singularity</a></h6><h6><a href="https://www.dwarkesh.com/i/161809078/transformative-ai">Transformative AI</a></h6><h6><a href="https://www.dwarkesh.com/i/161809078/explosive-economic-growth">Explosive economic growth</a></h6></blockquote><h6><a href="https://www.dwarkesh.com/i/161809078/alignment">Alignment</a></h6><blockquote><h6><a href="https://www.dwarkesh.com/i/161809078/reward-hacking">Reward hacking</a></h6><h6><a href="https://www.dwarkesh.com/i/161809078/takeover">Takeover</a></h6><h6><a href="https://www.dwarkesh.com/i/161809078/model-spec">Model spec</a></h6><h6><a href="https://www.dwarkesh.com/i/161809078/misuse">Misuse</a></h6></blockquote><h6><a href="https://www.dwarkesh.com/i/161809078/other">Other</a></h6><blockquote><h6><a href="https://www.dwarkesh.com/i/161809078/geopolitics">Geopolitics</a></h6><h6><a href="https://www.dwarkesh.com/i/161809078/epistemics">Epistemics</a></h6></blockquote><h2><strong>Capabilities</strong></h2><h3><strong>Agency</strong></h3><ul><li><p>Why don't we have reliable agents yet?</p></li><li><p>Is agency training just a veneer of some MCTS-like scaffolding on the knowledge &amp; intuition that pre-training gives you? Or is it much more difficult to develop?</p><ul><li><p>Here&#8217;s a case for agency being difficult to develop: <a href="https://epoch.ai/gradient-updates/movarec-s-paradox">Morevac&#8217;s Paradox</a>. Evolution has been optimizing us for hundreds of million of years for being able to act like a coherent goal seeking agent even in the face of super dynamic environments, whereas evolution has spent at most hundreds of thousands of years optimizing us for language skills and abstract reasoning. So it's not that surprising that we got expert-level AI mathematicians before AIs that can zero-shot video games made for 10-year-olds. Replicating capabilities derived from a billion years of evolutionary optimization might take much longer than replicating skills contingent on a hundred thousand years of optimization.</p><ul><li><p>The rebuttal to Morevac&#8217;s Paradox: the capabilities AIs are getting first have nothing to do with their recency in the evolutionary record and everything to do with how much relevant training data exists. Language and coding arrived first not because evolution only recently optimized us for reasoning but because we&#8217;ve got the fucking Internet and Github. Unitree robots are really good at walking around, despite the fact that evolution has spent over a quarter billion years teaching us locomotion - and this has everything to do with it being easy to get more data relevant to walking around using simulation.</p></li></ul></li></ul></li><li><p>What will be different about multi-agent systems?</p><ul><li><p>How much of a parallelization penalty will there be? Instead of one instance seeing and considering the whole context, you're breaking apart the problem to multiple workers.</p></li></ul></li><li><p>What is the explanation for why the length of task an AI can do doubles over consistent intervals?</p><ul><li><p>How will &#8220;<a href="https://x.com/METR_Evals/status/1902384481111322929">Moore&#8217;s Law for AI agents</a>&#8221; generalize to non-coding tasks like video editing, playing new video games, or coordinating logistics for a happy hour?</p></li><li><p>Would we get any intuition pumps about what superintelligence might look like by considering what it would mean to have a horizon length that is 10x as large as humans?</p></li></ul></li></ul><h3><strong>RL</strong></h3><ul><li><p>Dario said in <a href="https://www.darioamodei.com/post/on-deepseek-and-export-controls">his recent blog post on export controls</a> that labs are only spending on the order of $1M on RL - why? You&#8217;re spending hundreds of million on the base model. <strong>If RL training is such a big complement to pre-training, why not spend a similar amount of compute on it?</strong></p><ul><li><p>I keep hearing that the big bottleneck for RL is the amount of environments that we have built so far. I don't really understand what this means. What exactly does it take to build a new RL environment? Presumably building complex, realistic, hard to reward-hack challenges? <br>Is there a specific reason this is very hard (other than everything in this fallen world being harder than you might naively anticipate)?</p></li><li><p>Also you need smooth reward landscapes that allows AI to be rewarded for incremental improvements rather than getting stuck at 0. Smarter AIs have better priors about &#8220;what&#8217;s a reasonable thing to do when stuck&#8221;, which allow them to learn even from environments with sparser rewards.</p></li><li><p>By when will RL be the dominant workload in training? By when will most <a href="https://huggingface.co/learn/deep-rl-course/en/unitbonus3/offline-online">RL be online</a>?</p></li></ul></li><li><p>How sample-efficient is RL fine-tuning?</p><ul><li><p>For what kinds of skills is it especially effective? What skills are just hard to instill into the model, even if you have the appropriate data?</p></li><li><p>Even if agentic RL is sample-efficient, doesn&#8217;t it take way more compute per &#8216;sample&#8217; than RLHF-type training? <strong>As horizon lengths increase, your rollout has to become longer. The AI needs to do two hours worth of agentic computer use tasks before we can even see if it did it right. And if this is correct, will the pace of AI progress slow down?</strong></p><ul><li><p>Will this incentivize using smaller pre-trained models to do RL training? Would that allow more entrants to compete for the next tier of capabilities?</p></li></ul></li></ul></li><li><p>How much extra capability does the longer chain-of-thought get you (as opposed to just doing RL in the first place)?</p><ul><li><p>I don&#8217;t really understand why test-time compute scaling on things like the ARC-AGI benchmark keep giving marginal return (even if on a log scale). I get why thinking a bit longer might be helpful. But o3-high (which as of writing has the highest score)<a href="https://x.com/tobyordoxford/status/1907381570161315900"> wrote 43 million words</a> per task. What the hell did it figure out with its 42nd millionth word? Why do the benchmarks keep improving even up till that point?</p></li></ul></li><li><p>RL<a href="https://openreview.net/pdf?id=OwhVWNOBcz"> potentially</a> just upweights 10 tokens worth of MCTS-like scaffolding in a model's thinking (words like &#8220;wait&#8221;, &#8220;let&#8217;s backtrack&#8221;). This explains why reasoning models can be easily distilled - finding these basic techniques in thought space might take a while, but their payload size is trivial.</p><ul><li><p>So first question: is this actually correct?</p></li><li><p>How far can you distill reasoning and chain of thought? Will the models 6 months from now be able to do by instinct the kinds of math and coding that current models need to do large amounts of inference scaling for?</p></li></ul></li><li><p>How much transfer learning is there in reinforcement learning from verifiable reward?</p></li><li><p>Can RL work in non verifiable domains? Can you set up some kind of &#8216;peer review&#8217; with different agents/personas doing subjective critique?</p></li></ul><h3><strong>Idiot savants</strong></h3><ul><li><p>What is the answer to this question I asked Dario over a year ago? <em>As a scientist yourself, what should we make of the fact that despite having basically every known fact about the world memorized, these models haven&#8217;t, as far as I know, made a single new discovery? Even a moderately intelligent person who has so much stuff memorized would make all kinds of new connections (connect fact x, and y, and the logical implication is new discovery z).</em></p><ul><li><p>People have proposed all sorts of answers to this. For example, Scott Alexander wrote,<br>&#8220;Humans also aren't logically omniscient. My favorite example of this is etymology. Did you know that &#8216;vacation&#8217; comes from literally vacating the cities? Or that a celebrity is a person who is celebrated? Or that &#8216;dream&#8217; and &#8216;trauma&#8217; come from the same root? These are all kind of obvious when you think about them, but I never noticed before reading etymology sites.I think you don't make these connections until you have both concepts in attention at the same time, and the combinatorial explosion there means you've got to go at the same slow rate as all previous progress.&#8221;</p><ul><li><p>I agree that humans lack some godlike logical omniscience about the combinatorial consequences of all their knowledge. But there's plenty of examples of humans finding these kinds of important connections between fields, despite their much more limited world knowledge (see for instance<a href="https://x.com/dwarkesh_sp/status/1727018978420433286"> these examples</a>). I don't think Scott's argument explains why we have many examples of humans doing this, but none with AIs. And it's actually really funny that Scott is making this argument, because one of the things I love about his blog is that he <em>personally</em> has found countless numbers of these intellectually fascinating connections between fields. Where are the LLMs that have done this?</p></li></ul></li><li><p>Another argument I&#8217;ve seen: Eric Schmidt (no not that one)<a href="https://x.com/edavidds/status/1896443695312032195"> writes</a>, &#8220;I think I just don&#8217;t agree with your premise that there&#8217;s such low hanging fruit out there in the internet text corpus that hasn&#8217;t already been grabbed by smart, widely-read humans.&#8221;</p><ul><li><p>Given that the number of potential connections increases as O(N^2) with the amount of knowledge we have as species, and that the amount of knowledge is itself growing at least linearly, I find it implausible that humans have exhausted this combinatorial overhang.</p></li></ul></li><li><p>There is a &#8220;<a href="https://gwern.net/modus">one man's modus ponens is another man's modus tollens</a>&#8221; thing going on here. One way to interpret my original question is from an AI-skeptical point of view: The fact that LLMs aren't more powerful than humans despite their in-principle advantages over us suggests that they're not true AGI. But there's another interpretation of the same question - an interpretation which supports a more FOOMy vibe: Given the in-principle advantages LLMs have over us (in this case, as a result of their immense knowledge, but there are<a href="https://www.dwarkesh.com/p/ai-firm"> many others</a>), once they actually do become AGIs, they'll fucking dominate.</p></li></ul></li><li><p>Related to the question of LLMs knowing so much shit: It seems like knowledge is really cheap to store. Wikitext is less than 5 MB. So why do we humans forget so much? Why do our brains conspire so hard against acquiring new facts (which in raw bits cost basically nothing) that we have to come up with incredibly hacky systems like spaced repetition to keep knowledge around?</p><ul><li><p>One suggestive pattern, copy pasting from<a href="https://www.dwarkesh.com/p/symbolic-species"> my review of Terrence Deacon&#8217;s book, </a><em><a href="https://www.dwarkesh.com/p/symbolic-species">The Symbolic Species</a></em>: &#8220;Childhood amnesia (where you can&#8217;t remember early parts of your life) is the result of the learning process kids use, where they prune and infer a lot more, allowing them to see the forrest for the trees. On the opposite end of the spectrum are LLMs, which can remember entire passages of Wikipedia text verbatim but will flounder when you give them a new tic tac toe puzzle.There's something super interesting here where humans learn best at a part of their lives (childhood) whose actual details they completely forget, adults still learn really well but have terrible memory about the particulars of the things they read or watch, and LLMs can memorize arbitrary details about text that no human could but are currently pretty bad at generalization. It&#8217;s really fascinating that this memorization-generalization spectrum exists.&#8221;</p></li></ul></li><li><p>Can you have a 'superhuman AI scientist' before you get human level learning efficiency? (Currently, models take orders of magnitude more data that humans to learn equivalent skills, even ones they perform at 99th percentile level). </p><ul><li><p>My take is that creativity and learning efficiency are basically the same thing. The kind of thing Einstein did - generalizing from a few gnarly thought experiments and murky observations - is in some sense just extreme learning efficiency, right? Makes me wonder whether low learning efficiency is the answer to the question, 'Why haven't LLMs haven't made new discoveries despite having so much knowledge memorized'?</p></li></ul></li></ul><h3><strong>New training techniques</strong></h3><ul><li><p>I&#8217;m confused why all the labs have ended up making models that are so similar. Everyone is making &#8220;thinking&#8221; models. Has everybody just been trying a bunch of different shit, but this is the only thing that works? Or are they just copying each other, but in fact there&#8217;s a bunch of equally promising tangential research directions that no one is pursuing?</p></li><li><p>Many people have pointed out that there&#8217;s some missing middle between pre-training and in-context learning. Pre-training gives you some base of general understanding, something like a human skimming every textbook ever written; in-context learning is pure short term memory, discarded after every use. Are we likely to see a new training regime closer to dynamic evaluation, where you update your weights by meditating on feedback, or writing synthetic problems for practice?</p><ul><li><p>Would the procedure for dynamically learning new skills have to be bespoke for each application (if you need it to be good at call center workflows, you need to make a bespoke conversation trajectory unrolling environment) or can you come up with some general procedure for upskilling?</p></li></ul></li><li><p>One idea I&#8217;ve heard for building long horizon strategizing and coherency is to train AI systems of<a href="https://store.steampowered.com/tags/en/Text-Based/"> text based strategy games</a>. Has somebody already tried this and it didn&#8217;t work? Or it hasn&#8217;t even been tried in the first place?</p></li><li><p>What would a good benchmark for meta learning and sample efficient learning look like?</p><ul><li><p>If in context sample efficiency is high enough, is it okay if training sample efficiency sucks?</p></li></ul></li></ul><h3><strong>Pre-training</strong></h3><ul><li><p>Is pre-training actually dead?</p><ul><li><p>The compute required to train a GPT-4 level model has been declining in cost at the astonishing rate of 10x per year. I thought the whole point of these effective compute multipliers is that we could train GPT-5 at GPT- 4 cost. <strong>From the outside at least, it seems like these astonishing compute multipliers are only making existing capabilities cheaper to serve, not enabling the next generation of more powerful models to arrive much sooner. Rumors are that all the labs have been struggling to crack the next OOM of scaling. What&#8217;s going on? </strong>Data is running out? Maybe the engineering for much larger training runs gets exponentially harder? Or maybe so called algorithmic &#8216;compute multipliers&#8217; don&#8217;t give an equivalent multiplicative boost at different levels of scale?</p></li></ul></li><li><p>A couple months ago at ICML, Ilya compared the pre-training data corpus to fossil fuels - a limited resource which will rapidly be exhausted. This raises the question: what if this data corpus wasn&#8217;t limited? Would pre-training still be netting amazing new capabilities? Is next token prediction an innately fucked training objective, or did we just run out of data to let it cook?</p></li><li><p>Why are LLMs such mediocre writers despite having all the good writing in their training dataset?</p></li><li><p>Copy pasted <a href="https://www.lesswrong.com/posts/XiMRyQcEyKCryST8T/slowdown-after-2028-compute-rlvr-uncertainty-moe-data-wall">from Wei Dai</a>: &#8220;If the additional training data [after the 10s or trillions of tokens already used] is mostly low quality (AI labs must have used the highest quality data first?) or repetitive (contains no new ideas/knowledge), perplexity might go down but what is the LLM really learning?</p></li></ul><h2><strong>Economics</strong></h2><h3><strong>Early deployment</strong></h3><ul><li><p>Why aren't all the center workers getting laid off yet? It&#8217;s the first thing that should go. Should we take it as some signal that human jobs are just way harder to automate than you might naively think?</p></li><li><p>How will the difference between average S&amp;P 500 returns and median S&amp;P 500 company returns change over time?</p><ul><li><p>In other words, will we live in the world where Nvidia, Microsoft, Meta, and Google become worth $10T, but everything else goes to 0, or whether broad deployment happens fast enough that McDonalds, JP Morgan Chase, etc. become much more productive at the same rate that AI becomes more powerful.</p></li></ul></li><li><p>Will the models of 2026 and 2027 still best be thought of as time-shares of intelligence, or will they be integrated into companies and workflows such that individual copies are meaningfully distinct?</p><ul><li><p>Honestly, I think time-shares of intelligence are still underrated. A new AI employee can just read every single doc in your company's Drive and every single line of code in your company's codebase within minutes. This means that scaling up your company or an AI application would be way less effortful than scaling up a human department.</p></li></ul></li><li><p>What is the industrial-scale use case of AI? Between 1859 (when Drake first discovered oil in Pennsylvania), to 1908 (when Henry Ford invented the modern automobile), the main use for crude was as kerosene for lighting. What is the ultimate industrial-scale equivalent use case for AI?</p></li><li><p>How transformative would the AIs of today (March 2025) be even if AI progress stopped here?</p><ul><li><p>I've personally become more bearish about the economic value of current systems after using them to build miniapps for my podcast: I can't give them feedback that improves their scope or performance over time, and they can&#8217;t deal with unanticipated messy details.</p><ul><li><p>But then again, the first personal computers of the 1980s weren't especially useful either. They were mostly used by a few hobbyists. They had anemic memories and processing power, and there just wasn't a global network of applications to make them useful yet.</p></li><li><p>Another way to ask this question: imagine that you plopped down a steam engine in a hamlet from 1500. What would they do with it? Nothing! You need complementary technologies. There weren&#8217;t perfect steam engine shaped holes in these hamlets; similarly, there aren&#8217;t many LLM-shaped holes in today&#8217;s world</p></li></ul></li></ul></li></ul><h3><strong>Coding and remote work</strong></h3><ul><li><p>People like Dario say that in a year, 90%+ of code will be written by AIs. In some sense, compilers automated 90%+ of code writing. But I think compilers are a smaller deal than what these people imagine AI coders will be in a year. So what magnitude of actual productivity improvement to software engineering do they expect from AI? 2x? 10x? 100x?</p></li><li><p>If you get a 100x increase in software productivity, what kinds of things could we build in the world that we can&#8217;t build now?</p><ul><li><p>Here&#8217;s why I think this is such an interesting question. The economic historian Robert Allen <a href="https://www.amazon.com/Industrial-Revolution-Perspective-Approaches-Economic/dp/0521687853">thinks</a> cheap energy is a big part of the reason the Industrial Revolution first happened in Britain. The first steam engine (the Newcomen engine) was super inefficient. It worked not by pushing the piston directly with the steam but by relying on the steam condensing to pull the piston in. The only place where it made sense to even try this design was plentiful coal mines in England. This allowed Britain to get up the learning curve in mechanization, such that they could design devices which were viable with much less energy. The tl;dr here is that <strong>you need some initial hand-holding - some cold start - in order to get up the learning curve in a new technology. And because of cheap coal, Britain had it for the Industrial Revolution</strong>. <strong>Does effectively free software cause some kind of cold start to some future technological trend?</strong></p></li></ul></li><li><p>It seems like some AI labs (OpenAI, Meta) are racing towards being first to a billion user AI assistants, whereas others (Anthropic, maybe GDM?) are racing towards the fully autonomous software engineer. Who&#8217;s right? What are the marginal returns to more software (as compared to owning another Facebook- or Whatsapp-size social network)? If you believe in a software driven intelligence explosion, then that places a pretty high premium on the value of more software.</p><ul><li><p>The internet and mobile revolutions were largely driven by digital advertising &#8211; a relatively small slice of overall global GDP. If AI truly can automate all labor, how soon before AI revenues absolutely gobble up all the social/search/ads revenue of big tech?</p></li></ul></li><li><p>How much of the value of AI requires embodied robots (versus just digital LLMs)?</p></li></ul><h3><strong>Open source</strong></h3><ul><li><p>Does inference scaling mean that open-weight models don&#8217;t decentralize either the benefits or the risks of AI as much as you might naively think (more<a href="https://www.tobyord.com/writing/inference-scaling-reshapes-ai-governance"> here</a>)?</p></li><li><p>What does an intelligence explosion with open-weight models at the frontier look like?</p><ul><li><p>Is there any plausible way in which this could happen? Wouldn&#8217;t it require publishing the weights every week/day/hour (depending on speed of intelligence explosion)?</p></li><li><p>If this happened, would it be good?</p></li></ul></li><li><p><strong>What are the implications of the &#8220;panspermia&#8221; timeline, where multiple countries develop AGI from the same initial seed (e.g., by stealing model weights or building on open weights models)?</strong></p><ul><li><p>Would this shared foundation matter more if future advances come from <a href="https://www.alignmentforum.org/posts/vhfATmAoJcN8RqGg6/a-guide-to-iterated-amplification-and-debate">iterated amplification and distillation</a> rather than fresh training runs?</p></li><li><p>Does this increase AI takeover risk, since all the superintelligences start from the same subconscious &#8216;id&#8217;?</p></li><li><p>Does it increase the global impact of &#8216;Western&#8217; values which are embedded in those stolen weights?</p></li></ul></li></ul><h3><strong>Model training and value capture</strong></h3><ul><li><p>How are returns to training bigger models scaling? If a lab like OpenAI or DeepMind doubles its training compute budget for its <em>next</em> flagship model compared to the last one, are they getting more than 2x the revenue from it?</p><ul><li><p>It <a href="https://chatgpt.com/share/68053907-0d68-800e-ab46-c3f8734ad071">seems</a> like the ratio of revenue to training cost is increasing over time, plausibly because so much of the value of new skills is deferred to when AIs are fully reliable and complemented with a wider range of capabilities</p></li></ul></li></ul><ul><li><p>If progress in AI stalls, do foundation model companies inevitably get commoditized? Is there any moat other than staying 6+ months ahead (and in the extreme scenario beating everyone else to the intelligence explosion)? If model companies fail to differentiate, where does the value get captured?</p><ul><li><p>One answer might be the hyperscalers who control the datacenter compute, and whose complement (the models) just got commodified. But datacenter compute itself doesn&#8217;t seem that differentiated (so much so that the hyperscalers seem to be able to easily contract it out to third parties like CoreWeave). So maybe the lion&#8217;s share of value goes to the people making the components that go into chip production: 1) wafer production (TSMC), 2) advanced packaging (TSMC&#8217;s CoWoS), and 3) high bandwidth memory (SK Hynix)</p></li></ul></li><li><p>Many previous attempts to make AI applications based on scaffolds and wrappers have been gobbled up by better foundation models (which can just "scaffold" themselves). Will that keep being the case?</p></li><li><p>It's interesting to me that some of the best and most widely used applications of foundation models have come from the labs themselves (Deep Research, Claude Code, Notebook LM), even though it's not clear that you needed access to the weights in order to build them. Why is this? Maybe you do need access to the weights of frontier models, and the fine tuning APIs or open source models aren&#8217;t enough? Or maybe you gotta &#8216;feel the AGI&#8217; as strongly as those inside the labs do?</p></li><li><p>How much does being at the bleeding edge of AI capabilities matter? Is there any point in competing in the model race if you have no plan to get to the top? Or is there a viable business strategy based on being 6 months behind but following fast?</p></li></ul><h3><strong>Investment</strong></h3><ul><li><p>Hyperscaler AI capex is getting pretty big - approaching $100B a year for some of the big dogs. If there is a lull in AI capabilities (e.g., inference scaling doesn't end up being that useful or easy to scale, and it takes a couple of more years to make another big breakthrough), what will happen? Will fidgety CFOs force a firesale of compute contracts?</p></li><li><p>On the other hand, if AI does end up being as economically remunerative as "AGI" implies, how fast could the hyperscalers ramp up their investment? What would it take for them to invest more than their annual free cash flow (high 10s of Bs for Microsoft, Google, Amazon)? And how would they raise this money?</p></li></ul><h3><strong>Hardware</strong></h3><ul><li><p>Will the desire to avoid the 70%+ Jensen tax inevitably drive more in-house ASIC development?</p><ul><li><p>Will this drive greater diversity in model architectures and training techniques?</p></li></ul></li><li><p>How well will distributed training work?</p><ul><li><p>How easily can datacenters that have been built up assuming a pretraining focus be repurposed towards RL?</p></li><li><p>Will future training just look a lot like inference? There's not much difference between inference workloads and RL, since an agent has to go away for a while, try to solve a problem, and come back with the results.</p></li><li><p>Can RL rollouts just be totally <a href="https://x.com/PrimeIntellect/status/1912266266137764307">decentralized</a>?</p></li></ul></li><li><p>How much do hardware tariffs, energy costs, and geopolitical considerations influence where we build the next generation of AI data centers?</p><ul><li><p>If significant tariffs were imposed on datacenter components, would that shift planned buildouts towards Europe or Asia?</p></li><li><p>How much does network latency and proximity to customers matter when planning multi-billion dollar, decade-long infrastructure investments?</p></li></ul></li><li><p>How can we build hardware mechanisms that would help you prove to others what you&#8217;re using your compute for (and not using it for)?</p><ul><li><p>A lot of the stories about how the intelligence explosion goes well involve different actors instituting a mutual slow down where they refocus on alignment + greater monitoring. But such agreements require verifiability.</p></li></ul></li></ul><h2><strong>Post-AGI</strong></h2><h3><strong>Hive minds</strong></h3><ul><li><p>Will central planning actually work with AI?</p><ul><li><p>The case for:</p><ul><li><p>The central AI can have much higher bandwidth communication with the periphery. Today, sensing and deciding have to both happen at the front line. In the future, robot appendages of Big Brother could actually just run the whole military and economy.</p></li><li><p>The central planner can just directly learn from the experience of other AIs (think of a much more advanced version of Tesla FSD model learning from millions of driving samples).</p></li><li><p>Compute/intelligence can be much more centralized than it is today. Right now, Xi Jinping has the same 10^15 FLOPs as anyone else. Not so for the mega-inference-scaled dictators of the future.</p></li><li><p>In order to align incentives for humans, we have to use markets. But we can much more easily control the preferences of AIs.</p></li></ul></li><li><p>The case against:</p><ul><li><p>This vision assumes that only the central government is getting more complex and capable, while the rest of society stays similar. But AI deployment will lead to the whole economy becoming more complex. Apple circa 2025 might be able to centrally plan the economy of Babylon. But it wouldn&#8217;t be able to plan America circa 2025.</p></li></ul></li></ul></li><li><p>What is<a href="https://www.dwarkesh.com/p/ai-firm"> my blog post about fully automated firms</a> missing?</p></li><li><p><strong>How soon after AGI do you get hive minds, fully automated firms, and other crazy powerful and unique products of<a href="https://www.dwarkesh.com/p/ai-firm"> AIs&#8217; unique advantages in cultural learning/coordination</a> (namely, that AIs can copy, merge, distill, and scale themselves)?</strong></p><ul><li><p>An analogy is that the first AGIs will be like humans 200,000 years ago: yes, they have many key advantages, but the things that make us so dominant today - state capacity, joint stock corporations, fossil-fueled civilization - took eons of cultural evolution, population growth, technological upgrading.</p></li></ul></li><li><p>Today, the world economy runs on trade; complex, mutually-dependent supply chains; specialization; and decentralized knowledge. None of this looks like a singleton. Will all this fundamentally change once we have AGI? How about ASI?</p><ul><li><p>Or maybe a global system of interconnected markets, delegated decision-making, and mutual interdependence was the singleton all along?</p></li></ul></li></ul><h3><strong>Software only singularity</strong></h3><ul><li><p>How does the probability of an intelligence explosion change based on when we achieve AGI?</p><ul><li><p>AGI timelines tell you a lot about the nature of intelligence itself. If it takes decades to build AGI (rather than 3 more years of koding-koding-koding), then you&#8217;ve learned that transfer learning isn&#8217;t that powerful. You&#8217;ve learned that current systems aren&#8217;t simply &#8220;hobbled&#8221; little AGIs - rather, they will seem in retrospect more like AlphaZero - an amazing early demonstration of the possibility of AGI, but not the thing itself.</p></li></ul></li><li><p>Could an AI competent at research engineering accelerate AI progress such that we make a BERT to GPT-4.5 size jump in just one year?</p><ul><li><p>Case for: <a href="https://www.dwarkesh.com/i/160495909/debating-intelligence-explosion">my podcast with Daniel and Scott about AI 2027</a>.</p></li><li><p>Case against: <a href="https://www.dwarkesh.com/i/161532341/intelligence-explosion">my podcast with Ege &amp; Tamay.</a></p></li><li><p>AI labs apparently run multiple pre-training teams, each with varying allocations of compute and number of researchers. Observing how progress differs across these configurations would yield valuable insights into the tradeoff between cognitive effort and compute scaling in AI progress&#8212;though labs are unlikely to share this data.</p></li></ul></li></ul><ul><li><p>Suppose there's a software-only singularity. Would the gap between AI labs narrow or widen during this takeoff?</p><ul><li><p>There's a clear reason to expect the gap to widen during an intelligence explosion: whichever lab is slightly ahead benefits from having smarter, more capable automated researchers, rapidly compounding its advantage.</p></li><li><p>Yet historically, we've seen the gap consistently narrow. In 2018, DeepMind was far far ahead of everyone else. Today, labs like DeepSeek can close the distance to within half a year of OpenAI.</p></li><li><p>However, this pattern might not persist. The two primary methods for labs catching up&#8212;poaching experienced talent and learning from leaks or public deployments&#8212;would lose relevance in a scenario where the best models remain internal, autonomously driving progress.</p></li><li><p>Still, if the intelligence explosion accelerates by simultaneously scaling compute resources, the leading lab might remain incentivized to deploy publicly, attract investment, and thereby continue indirectly facilitating catch-up.</p></li></ul></li><li><p>Currently, model training (and inference) costs for the same level of capability are declining 10x a year. If you extrapolate this trend, then it suggests that you can train GPT-4 level capabilities using 100 H100s by 2027. Does this imply that while eventually we'll be able to distill human-level intelligence into mosquito drones, the first AGIs will be a system that costs $100B and requires Montana-sized infrastructure to train, and uses inference scaling hacks that cost $10 per token (like some Yoda, slowly meditating on each syllable)?</p></li></ul><h3><strong>Transformative AI</strong></h3><ul><li><p>Some people imagine that once we achieve ASI, it rapidly invents crazy superweapons, new computing paradigms, fusion-powered space probes, etc. But this doesn&#8217;t seem to map on to how past new inventions and scientific discoveries were made. It seems like a general upgrading of your society's tech stack is more important than raw cognitive effort in a specific sector: you couldn't have discovered the structure of DNA without X-ray crystallography; and we didn&#8217;t figure out there was a Big Bang until we saw cosmic background microwave thanks to the radio astronomy techniques that were initially developed for communications during World War II. Does this suggest that we're not just going to have this technological explosion in the middle of the desert? In order to truly advance the state of technology, you basically need to upgrade the whole economy. And this points to a broad deployment before an R&amp;D explosion.</p><ul><li><p>To be clear, transformative AI might still happen super fast! The AIs are thinking very quickly, there&#8217;s billions of them, they&#8217;re much better at learning-by-doing, etc. But it&#8217;s simply a question of whether or not we&#8217;ll just straight shot nano-tech or dyson spheres and skip over all the boring everyday improvements in seemingly unrelated fields.</p></li></ul></li></ul><h3><strong>Explosive economic growth</strong></h3><ul><li><p>Will explosive growth happen?</p><ul><li><p>The basic argument for explosive growth: Total economic output = Total Factor Productivity (TFP) &#215; Labor &#215; Capital. Today, economic growth has a weak feedback loop: output can accumulate into more capital, but not into more (human) labor. With AI, however, capital (compute) also functions as labor, creating a stronger feedback loop. Automated economic output = TFP (rapidly rising due to automated research and accelerated learning-by-doing) &#215; Capital (previous output).</p></li><li><p>If you&#8217;re interested in other considerations, check out <a href="https://www.dwarkesh.com/p/tyler-cowen-4">my interview with Tyler</a> for the con case, and the one with <a href="https://www.dwarkesh.com/p/ege-tamay">Ege &amp; Tamay</a> for the pro case.</p></li></ul></li><li><p>Will explosive growth basically look like what has happened to China post Deng (double digit growth rates instigated by abundant skilled labor)?</p></li><li><p>Will economic growth be extensive or intensive? Or some secret third thing?</p><ul><li><p>If the former, then it&#8217;s possible that the effects of explosive economic growth will not be felt that widely: your cityscape will not look that different; your life might not even change that much. But out in the desert, somewhere far from your view, they're producing solar farms and robot factories which are worth 200% of the existing economy.</p></li></ul></li></ul><h2><strong>Alignment</strong></h2><h3><strong>Reward hacking</strong></h3><ul><li><p>&#8220;LLMs were aligned by default. Agents trained with reinforcement learning reward hack by default&#8221; (<a href="https://x.com/ben_j_todd/status/1913206574706725301">tweet here</a>): is this actually the correct framing?</p><ul><li><p>Base LLMs were also misaligned by default. People had to figure out good post-training (partly using RL) to solve this. There's obviously no reward hacking in pretraining, but it&#8217;s not clear that pretraining vs RL have such different 'alignment by default'.</p></li></ul></li><li><p>Are there any robust solutions to reward hacking? Or is reward hacking such an attractive basin in training that if <em>any</em> exploit exists in the environment, models will train to hack it?</p><ul><li><p>Can we solve reward hacking by training agents in many different kinds of unique environments? In order to succeed, they&#8217;d have to develop robust general skills that don't just involve finding the exploits in any one particular environment.</p></li></ul></li><li><p>Do the models know when they&#8217;re reward hacking?</p></li><li><p><strong>Are capabilities and alignment the same thing here? Does making models more useful require solving reward hacking?</strong></p><ul><li><p><strong>If this is the case, we might be living in the alignment-by-default world? It would be weird if we solve reward hacking well enough to make these models reliable general agents in every scenario except those involved in taking over the world.</strong></p></li></ul></li></ul><h3><strong>Takeover</strong></h3><ul><li><p>What&#8217;s the right way to think about potentially misaligned AIs deployed through the economy? Are they like disgruntled employees who dislike the CEO (bullish for humanity), or more like Cort&#233;s landing in the New World (bearish)?</p><ul><li><p>An optimistic framing: Think of AI as just another set of workers. CEOs typically aren&#8217;t the smartest people in their companies, nor do they grasp all technical nuances of how their instructions are executed by engineers or researchers. The same applies at the national level&#8212;Xi Jinping doesn't lose sleep over whether all 100 million CCP members genuinely love him, yet he remains in power. Plenty of employees actively dislike their CEOs but still inadvertently contribute positively to the company's objectives. Why should we assume misaligned AIs would behave differently?</p></li><li><p>A pessimistic framing: history is littered with examples of violent takeovers enabled by moderate technological advantages. Geneticist <a href="https://www.youtube.com/watch?v=Uj6skZIxPuI">David Reich, on my podcast</a>, described human history as repetitive waves of violent expansion, where a technologically or organizationally superior group wipes out most people across entire continents. Consider Hern&#225;n Cort&#233;s, who conquered an empire of over 10 million people in just two years with fewer than 1000 soldiers (plus horses, steel, and Smallpox). Or the British East India Company, which took over a subcontinent of 150 million people with just a few thousand officers, aided by marginally superior artillery tactics and logistical innovations derived from European warfare&#8212;not some profound, galaxy-brain advantage. Perhaps AIs similarly wouldn&#8217;t need an overwhelming technological edge to decisively overpower humanity.</p></li></ul></li><li><p>If the AIs &#8216;takeover&#8217;, will we be able to see it coming? How wild and sci-fi will it be?</p><ul><li><p>Will it be like a gazelle being hunted by hunter-gatherers where it can clearly understand the strategic situation as it tries to run away from the pointy sticks?</p></li><li><p>Or will it be more like a deer peacefully grazing, unaware that it&#8217;s being watched through a scope 100 yards away, and a moment later&#8230; boom.</p></li></ul></li></ul><h3><strong>Model spec</strong></h3><ul><li><p>What should the model spec say?</p><ul><li><p>Alignment stories often assume we can get an AI to deeply internalize a foundational document that spells out its core values, directives, and ultimate authorities&#8212;something like, "Follow user instructions unless they're clearly dangerous; in disputes, defer to the judgment of Sam Altman or President Trump."</p></li><li><p>Think about how much the United States today is contingent on the specific quill strokes of Madison - what exactly does this comma mean, what did he mean by &#8216;general welfare&#8217; and &#8216;interstate commerce&#8217;?</p></li><li><p><strong>How do we get political opposition parties, allied governments, and other people in the world to fully grasp the enormous (temporary) leverage they currently have to influence this document?</strong></p></li></ul></li><li><p>If China develops AGI first, how do we make sure that their model spec doesn't just make Xi Jinping the god dictator forever? Do private Chinese companies and high-level members of the CCP have enough influence and understanding of the strategic situation to prevent such an outcome?</p></li></ul><h3><strong>Misuse</strong></h3><ul><li><p><strong>Is there an inherent tradeoff between preventing misuse and reducing the risk of a coup</strong> (either by AIs themselves or by humans using AI)? Or is this a false dichotomy?</p><ul><li><p>Preventing misuse looks a lot like making sure the most advanced capabilities aren't widely deployed because of the<a href="https://michaelnotebook.com/xriskbrief/index.html"> inherent dual-use nature of understanding reality</a>. Preventing takeover means making sure that we&#8217;re maximally open and transparent about frontier systems, and deploy them as widely as possible in order to prevent any one faction from monopolizing their benefits.</p></li></ul></li><li><p>Jailbreaking is becoming harder as models become smarter. This is pretty intuitive: a smart but ethical human assistant could tell if you&#8217;re asking him to help weaponize Smallpox or simply practicing for an organic chemistry exam. So as AIs get smarter, they should be able to tell the difference as well. Are we set here if we just put &#8220;don&#8217;t make bioweapons&#8221; in the model spec?</p></li><li><p>It seems like the whole misuse story is especially anchored on bioterrorism. The intuition pump is to imagine every person with a team of virology PhDs in her pocket. What actually would happen in this scenario?</p></li></ul><h2><strong>Other</strong></h2><h3><strong>Geopolitics</strong></h3><ul><li><p>I&#8217;m concerned about nationalization of AGI development for the reasons listed below. Where am I wrong?</p><ul><li><p>Reduces the salience of safety and alignment, in favor of whatever the administration + deep state at the time cares about.</p></li><li><p>Decreases the competence of the group overseeing the intelligence explosion, which is robustly bad, especially if you think alignment is difficult and subtle.</p></li><li><p>Increases the likelihood of dictatorship, with the President ending up as the ultimate decision maker in the spec or something.</p></li><li><p>Instigates an obvious arms race framing in China.</p></li><li><p>Increases the likelihood that the first AGIs are used to develop weapons, drones, and bioweapons instead of what&#8217;s needed for our glorious transhumanist future.</p></li></ul></li><li><p>How will the Chinese political system react to fully automated remote work, superhuman hackers, automated AI researchers etc.?</p><ul><li><p>How deep are their private markets? How willing are the big public funds to finance the next rung of scaling?</p></li><li><p>What might instigate a Chinese Manhattan Project for AGI?</p></li><li><p>Suppose tomorrow Xi wanted to prioritize building AGI. At an institutional level, what concretely could he do?</p></li><li><p>How does Chinese industrial espionage in other industries work? Suppose there's a hardware company that wants to learn how Apple does some procedure. Is there some government department they file their well-scoped snooping request with?</p></li></ul></li><li><p>Developing AI (and successfully deploying it at scale) is a huge industrial project. China is really good at industrial scale-up and state-directed investment. Does this give them a huge advantage in deploying AI?</p></li></ul><ul><li><p>If you&#8217;re the leader of a nation like India or Nigeria today, and you get AGI-pilled, what should you do?</p><ul><li><p>Honestly, it feels like a really tough position. The odds that you develop a frontier AI lab are pretty low. And you don&#8217;t have some crucial input into the semiconductor supply chain that will give you any leverage during crunch time.</p></li><li><p>If you have energy and property rights and friendly relations with the US and/or China, maybe you can become a datacenter hub?</p></li></ul></li><li><p>Does superintelligence actually give you &#8216;decisive strategic advantage&#8217;? Examples like Cortez or the East India Trading Company seem to imply that slight advantages in technology might be totally overwhelming in a military conflict. But perhaps modern weapons (nukes especially) are so destructive already that they provide sufficient deterrence against adversaries with smarter AIs launching their mosquito drone swarms and superhuman cyber attacks.</p></li></ul><h3><strong>Epistemics</strong></h3><ul><li><p>How wrong should we expect our conceptual handles around AGI to be?</p><ul><li><p>I've been reading <em>The House of Government </em>recently. It's a fascinating account of people involved in the Russian Revolution. There were many different factions of people who were disillusioned with the Czarist regime - the anarchists, the Mensheviks, Bolsheviks, the social revolutionaries, the Decembrists. They intensely debated the dichotomies which were most salient to them given their milieu.</p><ul><li><p><em>The &#8220;decisive battle&#8221; &#8230; <strong>covered all the usual points of disagreement: the &#8220;working class&#8221; versus &#8220;the people&#8221;; the &#8220;sober calculation&#8221; versus &#8220;great deeds and self-sacrifice&#8221;; &#8220;objectivism&#8221; versus &#8220;subjectivism&#8221;; and &#8220;universal laws of development&#8221; versus &#8220;Russia&#8217;s uniqueness.&#8221;</strong></em></p></li></ul></li><li><p>Yet none of them anticipated the considerations we now recognize to be far more relevant to economic development: dispersed knowledge, voluntary exchange, and entrepreneurial innovation.</p></li><li><p>I think about this whenever my Bay Area friends debate AGI - will there be a software-only singularity, adversarial misalignment, training gaming, explosive growth, etc, etc? Maybe the frameworks we're using and the questions we're asking are fundamentally misguided.</p></li></ul></li></ul><ul><li><p>Given this topic is so epistemically murky that someone smart can come up with a new consideration that alters your key conclusions, how much should you update on the most recent compelling story you&#8217;ve heard?</p><ul><li><p>Historically, the way we&#8217;ve dealt well with rapidly-evolving, uncertain processes is classical liberalism. Pushing for super tailored proposals based on reasoning your way to a specific trajectory (&#8220;We&#8217;re &lt;5 years to ASI, therefore &#8216;The Project&#8217;&#8221;) has a pretty bad track record.</p></li></ul></li></ul><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://www.dwarkesh.com/p/questions-about-ai?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">If you enjoyed this post, do share it with others who you think might enjoy it too.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.dwarkesh.com/p/questions-about-ai?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.dwarkesh.com/p/questions-about-ai?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><div><hr></div><p>If you enjoyed this blog post, you may enjoy my new book, <em><a href="https://www.stripe.press/scaling">The Scaling Era: An Oral History of AI, 2019-2025</a></em>. This book curates and organizes the highlights across <a href="https://www.youtube.com/@DwarkeshPatel">my podcast</a> episodes about AI with scientists, CEOs, economists, and philosophers.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2Mpy!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd171e5cb-6ada-4feb-99db-13403fdf658d_1600x1130.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2Mpy!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd171e5cb-6ada-4feb-99db-13403fdf658d_1600x1130.png 424w, https://substackcdn.com/image/fetch/$s_!2Mpy!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd171e5cb-6ada-4feb-99db-13403fdf658d_1600x1130.png 848w, https://substackcdn.com/image/fetch/$s_!2Mpy!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd171e5cb-6ada-4feb-99db-13403fdf658d_1600x1130.png 1272w, https://substackcdn.com/image/fetch/$s_!2Mpy!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd171e5cb-6ada-4feb-99db-13403fdf658d_1600x1130.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2Mpy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd171e5cb-6ada-4feb-99db-13403fdf658d_1600x1130.png" width="1456" height="1028" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d171e5cb-6ada-4feb-99db-13403fdf658d_1600x1130.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1028,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2Mpy!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd171e5cb-6ada-4feb-99db-13403fdf658d_1600x1130.png 424w, https://substackcdn.com/image/fetch/$s_!2Mpy!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd171e5cb-6ada-4feb-99db-13403fdf658d_1600x1130.png 848w, https://substackcdn.com/image/fetch/$s_!2Mpy!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd171e5cb-6ada-4feb-99db-13403fdf658d_1600x1130.png 1272w, https://substackcdn.com/image/fetch/$s_!2Mpy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd171e5cb-6ada-4feb-99db-13403fdf658d_1600x1130.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em>Thanks to Max Farrens for comments and editing. Thanks especially to Carl Shulman, but also Leopold Aschenbrenner, Tamay Besiroglu, Ege Erdil, Daniel Kokotjlo, Scott Alexander, Sholto Douglas, Adam D&#8217;Angelo, Paul Christiano, Craig Falls, Gwern Branwen, Dan Hendrycks, Andrej Karpathy, Toby Ord, Neel Nanda, and many others for conversations which helped inspire questions.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.dwarkesh.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe for future post!</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[What fully automated firms will look like]]></title><description><![CDATA[Everyone is sleeping on the *collective* advantages AIs will have, which have nothing to do with raw IQ  - they can be copied, distilled, merged, scaled, and evolved in ways humans simply can't.]]></description><link>https://www.dwarkesh.com/p/ai-firm</link><guid isPermaLink="false">https://www.dwarkesh.com/p/ai-firm</guid><dc:creator><![CDATA[Dwarkesh Patel]]></dc:creator><pubDate>Fri, 31 Jan 2025 16:02:21 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/81e63a95-d331-43d8-bdfb-a968a6ce213f_1792x1024.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Developed in collaboration with <a href="https://x.com/EgeErdil2">Ege Erdil</a> and <a href="https://x.com/tamaybes">Tamay Besiroglu</a>. Thanks to <a href="https://x.com/rebeccahiscott?lang=en">Rebecca Hiscott</a> and <a href="https://x.com/g_leech_">Gavin Leech</a> for editing and comments.</em></p><p><em>Epistemic status: Shooting-the-shit; 25% sure this roughly describes how firms of AGIs will actually work.</em></p><p>Even people who expect human-level AI soon are still seriously underestimating how different the world will look when we have it. Most people are anchoring on how smart they expect individual models to be. (i.e. they&#8217;re asking themselves &#8220;What would the world be like if everyone had a very smart assistant who could work 24/7?&#8221;.)</p><p>Everyone is sleeping on the <em>collective</em> advantages AIs will have, which have nothing to do with raw IQ but rather with the fact that they are digital&#8212;they can be copied, distilled, merged, scaled, and evolved in ways human simply can&#8217;t.</p><p>What would a fully automated company look like - with all the workers, all the managers as AIs? I claim that such AI firms will grow, coordinate, improve, and be selected-for at unprecedented speed.</p><p>This essay is not a prediction of what GPT-5 will be doing, nor about emulations of existing humans. Rather, I'm trying to imagine what the world will look like once we actually have AGIs - the descendants of LLMs that have gotten so good that they can do basically anything any human can do.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.dwarkesh.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.dwarkesh.com/subscribe?"><span>Subscribe now</span></a></p><h3><strong>Copy</strong></h3><p>Currently, firms are extremely bottlenecked in hiring and training talent. But if your talent is an AI, you can copy it a stupid number of times. What if Google had a million AI software engineers? Not untrained amorphous "workers," but the AGI equivalents of Jeff Dean and Noam Shazeer, with all their skills, judgment, and tacit knowledge intact.</p><p>This ability to turn capital into compute <em>and compute into equivalents of your top talent</em> is a fundamental transformation. Since you can amortize the training cost across thousands of copies, you could sensibly give these AIs ever-deeper expertise - PhDs in every relevant field, decades of business case studies, intimate knowledge of every system and codebase the company relies on.</p><p>The power of copying extends beyond individuals to entire teams. Small previously successful teams (think PayPal Mafia, early SpaceX, the <a href="https://en.wikipedia.org/wiki/Traitorous_eight">Traitorous Eight</a>) can be replicated to tackle a thousand different projects simultaneously. It's not just about replicating star individuals, but entire configurations of complementary skills that are known to work well together. The unit of replication becomes whatever <em>collection </em>of talent has proven most effective.</p><p>Copying will transform management even more radically than labor. It will enable a level of micromanagement that makes founder mode look quaint. Human Sundar simply doesn't have the bandwidth to directly oversee 200,000 employees, hundreds of products, and millions of customers. But AI Sundar&#8217;s bandwidth is capped only by the number of TPUs you give him to run on. All of Google&#8217;s <a href="https://archive.is/Fz4zx">30,000 middle managers</a> can be replaced with AI Sundar copies. <strong>Copies of AI Sundar can craft every product&#8217;s strategy, review every pull request, answer every customer service message, and handle all negotiations - everything flowing from a single coherent vision.</strong></p><p>There is no principal-agent problem wherein employees are optimizing for something other than Google&#8217;s bottom line, or simply lack the judgment needed to decide what matters most.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> A company of Google's scale can run much more as the product of a single mind&#8212;the articulation of one thesis&#8212;than is possible now.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a></p><h3><strong>Merge</strong></h3><p>Think about how limited a CEO's knowledge is today. How much does Sundar Pichai really know about what's happening across Google's vast empire? He gets filtered reports and dashboards, attends key meetings, and reads strategic summaries. But he can't possibly absorb the full context of every product launch, every customer interaction, every technical decision made across hundreds of teams. His mental model of Google is necessarily incomplete.</p><p>Now imagine mega-Sundar &#8211; the central AI that will direct our future AI firm. Just as Tesla's Full Self-Driving model can learn from the driving records of millions of drivers, mega-Sundar might learn from everything seen by the distilled Sundars - every customer conversation, every engineering decision, every market response.</p><p>Unlike Tesla&#8217;s FSD, this doesn&#8217;t have to be a naive process of gradient updating and averaging. Mega-Sundar will absorb knowledge far more efficiently &#8211; through explicit summaries, shared latent representations, or even surgical modification of the weights to encode specific insights.</p><p>The boundary between different AI instances starts to blur. Mega-Sundar will constantly be spawning specialized distilled copies and reabsorbing what they&#8217;ve learned on their own. Models will communicate directly through <a href="https://colala.berkeley.edu/papers/piantadosi2024why.pdf">latent representations</a>, similar to how the hundreds of different layers in a neural network like GPT-4 already interact.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> So, approximately no miscommunication, ever again. The relationship between mega-Sundar and its specialized copies will mirror what we're already seeing with techniques like speculative decoding &#8211; where a smaller model makes initial predictions that a larger model verifies and refines.</p><p>Merging will be a step change in how organizations can accumulate and apply knowledge. Humanity's great advantage has been social learning &#8211; our ability to pass knowledge across generations and build upon it. But human social learning has a terrible handicap: biological brains don't allow information to be copy-pasted. So you need to spend years (and in many cases decades) teaching people what they need to know in order to do their job. Look at how top achievers in field after field are getting <a href="https://anilv.com/age">older and older</a>, maybe because it takes longer to reach the frontier of accumulated knowledge. Or consider how <a href="https://www.nber.org/system/files/chapters/c7977/c7977.pdf">clustering talent</a> in cities and top firms produces such outsized benefits, simply because it enables slightly better knowledge flow between smart people.</p><p>Future AI firms will accelerate this cultural evolution through two key advantages: massive population size and perfect knowledge transfer. With millions of AGIs, automated firms get so many more opportunities to produce innovations and improvements, whether from lucky mistakes, deliberate experiments, de-novo inventions, or some combination. </p><p>As Joseph Henrich explains in <a href="https://www.amazon.com/WEIRDest-People-World-Psychologically-Particularly/dp/0374173222">The WEIRDest People in the World</a>,</p><blockquote><p>cumulative cultural evolution&#8212;including innovation&#8212;is fundamentally a social and cultural process that turns societies into collective brains. Human societies vary in their innovativeness due in large part to the differences in the fluidity with which information diffuses through a population of engaged minds and across generations</p></blockquote><p><a href="https://faculty.econ.ucdavis.edu/faculty/gclark/210a/readings/kremer1993.pdf">Historical data going back thousands of years</a> suggest that population size is the key input for how fast your society comes up with more ideas. AI firms will have population sizes that are orders of magnitude larger than today's biggest companies - and each AI will be able to perfectly mind meld with every other, from the bottom to the top of the org chart.</p><p>AI firms will look from the outside like a unified intelligence that can instantly propagate ideas across the organization, preserving their full fidelity and context. Every bit of tacit knowledge from millions of copies gets perfectly preserved, shared, and given due consideration.</p><h3><strong>Scale</strong></h3><p>The cost to have an AI take a given role will become just the amount of compute the AI consumes. This will change our understanding of which roles are scarce.</p><p>Future AI firms won&#8217;t be constrained by what's scarce or abundant in human skill distributions &#8211; they can optimize for whatever abilities are most valuable. Want Jeff Dean-level engineering talent? Cool: once you&#8217;ve got one, the marginal copy costs pennies. Need a thousand world-class researchers? Just spin them up. The limiting factor isn't finding or training rare talent &#8211; it's just compute.</p><p>So what becomes expensive in this world? Roles which justify massive amounts of <a href="https://huggingface.co/spaces/HuggingFaceH4/blogpost-scaling-test-time-compute">test- time compute</a>. The CEO function is perhaps the clearest example. Would it be worth it for Google to spend $100 billion annually on inference compute for mega-Sundar? Sure! Just consider what this buys you: millions of subjective hours of strategic planning, Monte Carlo simulations of different five-year trajectories, deep analysis of every line of code and technical system, and exhaustive scenario planning.</p><p>Imagine mega-Sundar contemplating: "How would the FTC respond if we acquired eBay to challenge Amazon? Let me simulate the next three years of market dynamics... Ah, I see the likely outcome. I have five minutes of datacenter time left &#8211; let me evaluate 1,000 alternative strategies."</p><p>The more valuable the decisions, the more compute you'll want to throw at them. A single strategic insight from mega-Sundar could be worth billions. An overlooked risk could cost tens of billions. However many billions Google should optimally spend on inference for mega-Sundar, it's certainly more than <a href="https://www.ft.com/content/666d8609-3d7b-4757-97ed-2a86fb99d914">one</a>.</p><h3><strong>Distillation</strong></h3><p>What might distilled copies of AI Sundar (or AI Jeff) be like? Obviously, it makes sense for them to be highly specialized, especially when you can amortize the cost of that domain specific knowledge across all copies. You can give each distilled data center operator a deep technical understanding of every component in the cluster, for example.</p><p>I suspect you&#8217;ll see a lot of specialization in function, tacit knowledge, and complex skills, because they seem expensive to sustain in terms of parameter count. But I think the different models might share a lot more factual knowledge than you might expect. It&#8217;s true that plumber-GPT doesn&#8217;t need to know much about the standard model in physics, nor does physicist-GPT need to know why the drain is leaking. But the cost of storing raw information is so unbelievably cheap (and it&#8217;s only decreasing) that Llama-7B already knows more about the standard model and leaky drains than any non-expert. If human-level intelligence is more than 1 trillion parameters, is it so much of an imposition to keep around what will, at the limit, be much less than 7 billion parameters to have <em>most known facts</em> right in your model? (Another helpful data point here is that &#8220;Good and Featured&#8221; Wikitext is <a href="https://huggingface.co/datasets/Salesforce/wikitext">less than 5 MB</a>. I don&#8217;t see why all future models&#8212;except the esoteric ones, the digital equivalent of tardigrades&#8212;wouldn&#8217;t at least have Wikitext down.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a></p><h3><strong>Evolve</strong></h3><p>The most profound difference between AI firms and human firms will be their evolvability. As Gwern Branwen <a href="https://gwern.net/backstop">observes</a>:</p><blockquote><p>Why do we not see exceptional corporations clone themselves and take over all market segments? Why don&#8217;t corporations evolve such that all corporations or businesses are now the hyper-efficient descendants of a single ur-corporation 50 years ago, all other corporations having gone extinct in bankruptcy or been acquired? Why is it so hard for corporations to keep their &#8220;culture&#8221; intact and retain their youthful lean efficiency, or, if avoiding &#8220;aging&#8221; is impossible, why [not] copy themselves or otherwise reproduce to create new corporations like themselves?</p></blockquote><p>His answer:</p><blockquote><p>Corporations certainly undergo selection for kinds of fitness, and do vary a lot. The problem seems to be that corporations cannot replicate themselves &#8230; Corporations are made of people, not interchangeable, easily copied widgets or strands of DNA .. The corporation may not even be able to &#8220;replicate&#8221; itself over time, leading to scleroticism and aging.</p></blockquote><p>The scale of difference between currently existing human firms and fully automated firms will be like the gulf in complexity between prokaryotes and eukaryotes. <s>Prokaryotes like </s><a href="https://www.lesswrong.com/posts/wpRP44NT6kHaKNdnn/why-are-bacteria-so-simple"><s>bacteria are not only remarkably simple</s></a><s>, but have barely changed over their 3 billion year history</s><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a>. Whereas eukaryotes rapidly scaled up in complexity, and gave rise to all the other astonishing organisms with trillions of cells working together tightknit<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a></p><p>This evolvability is also the key difference between AI and human firms. As Gwern points out, human firms simply cannot replicate themselves effectively - they're made of people, not code that can be copied. They can't clone their culture, their institutional knowledge, or their operational excellence. AI firms can<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a>.</p><p>If you think human Elon is especially gifted at creating hardware companies, you simply can&#8217;t spin up 100 Elons, have them each take on a different vertical, and give them each $100 million in seed money. As much of a micromanager as Elon might be, he&#8217;s still limited by his single human form. But AI Elon can have copies of himself design the batteries, be the car mechanic at the dealership, and so on. And if Elon isn&#8217;t the best person for the job, the person who <em>is</em> can also be replicated, to create the template for a new descendant organization.</p><h3><strong>Takeover</strong></h3><p>So then the question becomes: If you can create Mr. Meeseeks for any task you need, why would you ever pay some markup for another firm, when you can just replicate them internally instead? Why would there even be other firms? Would the first firm that can figure out how to automate everything will just form a conglomerate that takes over the entire economy?</p><p>Ronald Coase&#8217;s <a href="https://en.wikipedia.org/wiki/Theory_of_the_firm">theory of the firm</a> tells us that companies exist to reduce transaction costs (so that you don&#8217;t have to go rehire all your employees and rent a new office every morning on the free market). His theory states that the lower the intra-firm transaction costs, the larger the firms will grow. Five hundred years ago, it was practically impossible to coordinate knowledge work across thousands of people and dozens of offices. So you didn&#8217;t get very big firms. Now you can spin up an arbitrarily large Slack channel or HR database, so firms can get much bigger.</p><p>AI firms will lower transaction costs so much relative to human firms. It&#8217;s hard to beat shooting lossless latent representations to an exact copy of you for communication efficiency! So firms probably will become much larger than they are now.</p><p>But it&#8217;s not inevitable that this ends with one gigafirm which consumes the entire economy. As Gwern explains in his essay, any internal planning system needs to be grounded in some kind of outer "loss function" - a ground truth measure of success. In a market economy, this comes from profits and losses.</p><p>Internal planning can be much more efficient than market competition in the short run, but it needs to be constrained by some slower but unbiased outer feedback loop. A company that grows too large risks having its internal optimization diverge from market realities.</p><p>That said, the balance may shift as AI systems improve. As corporations become more "software-like" - with perfect replication of successful components and faster feedback loops - we may see much larger and more efficient firms than were previously possible.</p><p>The market continues to serve as the grounding outer loop. How does the firm convert trillions of tokens of data from customers, markets, news, etc every day into future plans, new products, and the like? Does the board make all the decisions politburo-style and use $10 billion dollars of inference to run Monte Carlo tree search on different one-year plans? Or do you run some kind of evolutionary process on different departments, giving them more capital, and <strong>compute/labor</strong> based on their performance?</p><p>These are all what we would today call &#8220;culture.&#8221; Markets facilitate an evolutionary process which selects not only goods and services, but the institutions that are best at turning the world into valuable goods and services. I think this will continue.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>&#8230;except for the biggie: the intensified principal-agent problem between the CEO-workers (who suddenly know everything) and the shareholders on the outside (who know as much as they know now, i.e. roughly nothing).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Since labor is trivial to copy and spin up, the value of intellectual property will go up. The essence of the firm basically becomes intellectual property. GM can poach as many Tesla engineers as it wants (or, in our hypothetical, clone someone with equivalent skills). But without intellectual theft, they can&#8217;t get the FSD model or the millions of hours of driving it was trained on. If firms no longer have a moat in labor, their moat will be this kind of industry-specific knowledge and data.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>An intriguing thought, in response to things like <a href="https://arxiv.org/html/2412.06769v1">Meta&#8217;s Continuous Chain of Thought </a>models, is that actually we will want to keep communication happening in discrete tokens, since this is a kind of autoencoding - a bottleneck which makes you usefully compress information into actual insight when you communicate. (Leaders have to force people to keep things simple, since in general people want to sound smart instead.) But these tokens don&#8217;t need to relate to natural language necessarily.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>There&#8217;s a biological analogy here, too. Every cell in your body stores each base pair of your 3GB DNA, despite the fact that only 10 to 20 percent of the protein coding regions are expressed in any particular cell, and only 1 percent of your DNA is protein coding in the first place, with much of the rest long thought of as &#8220;<a href="https://www.genomicseducation.hee.nhs.uk/genotes/knowledge-hub/non-coding-dna/">junk</a>&#8221;. Information is apparently just so cheap to store that there&#8217;s been little selective pressure against this redundancy and waste.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>The biologist Niko McCarty corrects me <a href="https://x.com/NikoMcCarty/status/1885504401223070074">here</a>! I think the general point still stands that endosymbiosis was a regime change in evolutionary complexity. </p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>Big changes in evolvability don&#8217;t have to involve too much complexity or mutation. Two reasons bacteria didn&#8217;t evolve much compared to eukaryotes is that they performed energy production on their surface which scales poorly with cell volume (compared to eukaryotes which solved this with mitochondria); and because they had a constraint on their genome size (because they compete with each other by replicating faster, and replication is mainly bottlenecked on DNA length). </p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>A <a href="https://are.berkeley.edu/~aprajit/DMM.pdf">study</a> by Bloom et al. showed that giving firms' management training increases their productivity by 17 percent. With AI firms, this could be a simple immediate upgrade closer to horizontal gene transfer - i.e. available to all copies, not just one particular treatment group.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[Notes on China]]></title><link>https://www.dwarkesh.com/p/notes-on-china</link><guid isPermaLink="false">https://www.dwarkesh.com/p/notes-on-china</guid><dc:creator><![CDATA[Dwarkesh Patel]]></dc:creator><pubDate>Fri, 27 Dec 2024 19:09:51 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbad23033-78a3-44ef-8484-a2314dc7121b_4032x3024.heic" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Last month, I spent 2 weeks in China - I visited Beijing, Chengdu, Emeishan, Chongqing, Shanghai, and Hangzhou.</p><p>I have no illusions about "understanding" China. I've only spent 2 weeks there. This trip is a beginning, not a capstone, of my curiosity about China. And I hope to share what I learn with you via future podcast episodes.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.dwarkesh.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.dwarkesh.com/subscribe?"><span>Subscribe now</span></a></p><h2>Scale</h2><p>It&#8217;s funny how China has basically the inverse problem as America. We subsidize demand and restrict supply. They subsidize supply and restrict demand. We can&#8217;t rebuild fallen bridges. They build bridges to nowhere. In the most desirable cities in this country, every random Victorian house and park bench is a historic site that can&#8217;t be disturbed. There, they&#8217;ll bulldoze a 500 year old temple to build an endless skyscraper complex that no one wants to live in.</p><p>My overwhelming first impression was: wow everything is so fucking big: the cities themselves, the train stations, the airports,  the towering and endless apartment complexes. Travel often teaches you things about a country which you honestly should have intuited even without visiting. Obviously, I knew that China is a big country, with over 1.4 billion people. But the stupendous scale of the biggest cities was impressed upon me only after I visited.</p><p>Even in Emeishan, a city of just half a million people (considered a quaint countryside town by Chinese standards), we found a Buddhist temple of comical scale - we'd enter what seemed like an impressively large compound, only to discover it was merely the entrance to an even grander structure right behind it. This pattern repeated 5 or 6 times, each subsequent building larger and more ornate than the last, like some kind of inverse nesting dolls. And the place had almost no other visitors!</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pY-M!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c4c80ed-b1ca-4782-b30a-6f4133a48bf0_4032x3024.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pY-M!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c4c80ed-b1ca-4782-b30a-6f4133a48bf0_4032x3024.heic 424w, https://substackcdn.com/image/fetch/$s_!pY-M!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c4c80ed-b1ca-4782-b30a-6f4133a48bf0_4032x3024.heic 848w, https://substackcdn.com/image/fetch/$s_!pY-M!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c4c80ed-b1ca-4782-b30a-6f4133a48bf0_4032x3024.heic 1272w, https://substackcdn.com/image/fetch/$s_!pY-M!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c4c80ed-b1ca-4782-b30a-6f4133a48bf0_4032x3024.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pY-M!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c4c80ed-b1ca-4782-b30a-6f4133a48bf0_4032x3024.heic" width="1456" height="1092" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7c4c80ed-b1ca-4782-b30a-6f4133a48bf0_4032x3024.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1092,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:3205969,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!pY-M!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c4c80ed-b1ca-4782-b30a-6f4133a48bf0_4032x3024.heic 424w, https://substackcdn.com/image/fetch/$s_!pY-M!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c4c80ed-b1ca-4782-b30a-6f4133a48bf0_4032x3024.heic 848w, https://substackcdn.com/image/fetch/$s_!pY-M!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c4c80ed-b1ca-4782-b30a-6f4133a48bf0_4032x3024.heic 1272w, https://substackcdn.com/image/fetch/$s_!pY-M!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c4c80ed-b1ca-4782-b30a-6f4133a48bf0_4032x3024.heic 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">This must have been structure number 2 or 3?</figcaption></figure></div><p>I asked a monk at the temple how they funded this massive site in a city of just half a million people. He told us that it was simply through donations. We probed further about how such an enormous project could have been financed by just ordinary people's contributions. He responded, "We've got a lot of supporters, dude", and changed the topic.</p><p>Chongqing is by far the coolest city I've ever visited. It's this insane cyberpunk multi-level metropolis of over 20 million people. I wouldn't know how to begin describing it, but there's a bunch of great YouTube <a href="https://youtu.be/IbnSuon2nrI?si=UuyLwwOpD3QWCgOs">videos</a> which will show you what I mean. I got a really nice nice 2-floor hotel room that overlooked two rivers and one of the most insane skylines in the world for 60 bucks - highly recommend visiting Chongqing if you get the chance.</p><p>In 1995, astronomers pointed Hubble at a seemingly empty patch of sky the size of a grain of sand held at arm's length. Instead of emptiness, the 10-day exposure revealed over 3,000 galaxies. Every speck of light in the image was an entire galaxy containing billions of stars. When I went atop the tallest building in Chongqing and looked out over the city, I thought about the Hubble image. You could zoom in on any direction you'd find - behind the fog and mist, beyond even perhaps the horizon - another skyscraper, each containing hundreds or thousands of people living or working.</p><div class="image-gallery-embed" data-attrs="{&quot;gallery&quot;:{&quot;images&quot;:[{&quot;type&quot;:&quot;image/jpeg&quot;,&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/12811026-40ed-4787-8447-67fa59919542_4032x3024.jpeg&quot;},{&quot;type&quot;:&quot;image/heic&quot;,&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5a4dd8f8-8fc5-4b04-93f1-d540170ac5fd_3024x4032.heic&quot;},{&quot;type&quot;:&quot;image/heic&quot;,&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6bb643a9-f5ca-43c8-b7d8-50ad7518b9f9_4032x3024.heic&quot;}],&quot;caption&quot;:&quot;&quot;,&quot;alt&quot;:&quot;&quot;,&quot;staticGalleryImage&quot;:{&quot;type&quot;:&quot;image/png&quot;,&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e670a756-28e9-4949-bde4-32b4232d3c8a_1456x474.png&quot;}},&quot;isEditorNode&quot;:true}"></div><p>We took a 12-hour village train from Chongqing to Shanghai. I'm embarrassed to say that my only experience with the actual countryside was via the windows of this train. Still, the sights were quite interesting. We saw again and again small paddy farms surrounding a handful of 5-10 story skyscraper, plopped in the seeming middle of nowhere. Even in the countryside, many people lived in large buildings instead of their own small homes. I couldn't hop off the train and confirm, but I saw many towns that looked quite ghostly - no actually visible people anywhere.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Xf5N!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbad23033-78a3-44ef-8484-a2314dc7121b_4032x3024.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Xf5N!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbad23033-78a3-44ef-8484-a2314dc7121b_4032x3024.heic 424w, https://substackcdn.com/image/fetch/$s_!Xf5N!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbad23033-78a3-44ef-8484-a2314dc7121b_4032x3024.heic 848w, https://substackcdn.com/image/fetch/$s_!Xf5N!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbad23033-78a3-44ef-8484-a2314dc7121b_4032x3024.heic 1272w, https://substackcdn.com/image/fetch/$s_!Xf5N!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbad23033-78a3-44ef-8484-a2314dc7121b_4032x3024.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Xf5N!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbad23033-78a3-44ef-8484-a2314dc7121b_4032x3024.heic" width="1456" height="1092" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bad23033-78a3-44ef-8484-a2314dc7121b_4032x3024.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1092,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2287698,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Xf5N!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbad23033-78a3-44ef-8484-a2314dc7121b_4032x3024.heic 424w, https://substackcdn.com/image/fetch/$s_!Xf5N!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbad23033-78a3-44ef-8484-a2314dc7121b_4032x3024.heic 848w, https://substackcdn.com/image/fetch/$s_!Xf5N!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbad23033-78a3-44ef-8484-a2314dc7121b_4032x3024.heic 1272w, https://substackcdn.com/image/fetch/$s_!Xf5N!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbad23033-78a3-44ef-8484-a2314dc7121b_4032x3024.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Outside of Beijing and Shanghai (and sometimes even within), you can tell that these skyscrapers were put up by a country with a GDP per capita of $10,000 (and potentially half or quarter than when many of these buildings went up). America and Europe put up a ton of beautiful buildings in the early 20th century when their GDP per capita was similar to China's - one could even argue that those older structures are more aesthetic than anything we're building today in the West. Not so in China. These endless rows of skyscrapers, put up in the construction frenzy of the last few decades, are ugly - boxes of mostly concrete with visible blight and discoloration all over them. If the great construction binge is indeed over, it'll be a shame that China's infrastructure was built out during a period of particularly uninspired architecture.</p><p>Beijing's urban design looks like something straight out of James Scott's "Seeing Like a State". The city is dominated by these enormous apartment complexes - blocks of 10 adjacent 30-story buildings demarcated by 8-lane roads. The government buildings follow the same pattern: huge structures divided by extremely wide boulevards. This layout seems designed partly for social control - during zero-COVID, authorities could lock down 10,000 people by simply guarding a few entrance gates. The wide roads would also make it easy to move military forces through the city. The only break from this pattern are the Hutongs, Beijing's old historic neighborhoods. But even these weren't spared completely - only a fraction survived Beijing's rapid modernization push. Dare I say that China is too YIMBY?</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!eBz0!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7918506-c257-4a24-91ea-2063c48800ed_3501x2334.avif" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!eBz0!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7918506-c257-4a24-91ea-2063c48800ed_3501x2334.avif 424w, https://substackcdn.com/image/fetch/$s_!eBz0!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7918506-c257-4a24-91ea-2063c48800ed_3501x2334.avif 848w, https://substackcdn.com/image/fetch/$s_!eBz0!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7918506-c257-4a24-91ea-2063c48800ed_3501x2334.avif 1272w, https://substackcdn.com/image/fetch/$s_!eBz0!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7918506-c257-4a24-91ea-2063c48800ed_3501x2334.avif 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!eBz0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7918506-c257-4a24-91ea-2063c48800ed_3501x2334.avif" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c7918506-c257-4a24-91ea-2063c48800ed_3501x2334.avif&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:989725,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/avif&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!eBz0!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7918506-c257-4a24-91ea-2063c48800ed_3501x2334.avif 424w, https://substackcdn.com/image/fetch/$s_!eBz0!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7918506-c257-4a24-91ea-2063c48800ed_3501x2334.avif 848w, https://substackcdn.com/image/fetch/$s_!eBz0!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7918506-c257-4a24-91ea-2063c48800ed_3501x2334.avif 1272w, https://substackcdn.com/image/fetch/$s_!eBz0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7918506-c257-4a24-91ea-2063c48800ed_3501x2334.avif 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Vibes</h2><p>I got quite mixed messages about the state of public opinion in China. This is to be expected in a society where you can't establish common knowledge. One person told me that the new generation is quite nationalist, unlike the older reform generation which personally experienced the catastrophes of Mao and the tangible benefits of liberalization. He made the rather insightful point that this tilt in Chinese public opinion increasingly gives lie to the American talking point, "We're against the CCP, not the Chinese people." In fact, he went on to say that the current regime is way more liberal than what would result from an election in China.</p><p>Another person told me that these Chinese nationalists were only a vocal minority, similar to the wokes in America circa 2020. While they make up only about 10% of the population, they aggressively shout down others on Weibo (China's Twitter equivalent). Most people find them annoying but feel uncomfortable confronting them directly. This matches what a student who graduated from a top university there told me - the vast majority of his classmates are simply apolitical. And in our own interactions with locals, we saw little evidence of widespread nationalism. In fact, when my Chinese-speaking trip mate would mention he was from the UK to taxi drivers, they would often respond enthusiastically: "Oh wonderful, we love the UK!"</p><p>There were very few foreigners. In Beijing I might have seen half a dozen cumulatively across entire seas of people. In Chengdu and Chongqing, I barely remember seeing any (which is a real shame, because Chongqing is a truly incredible tourist destination). So rare apparently are tourists that in Chengdu and Chongqing we got asked for selfies many times.</p><p>Outside of Shanghai, almost nobody spoke English. If I went again, I would definitely try to crash course some basic Chinese beforehand. This language barrier did led to some interesting encounters. At a park in Chengdu, an old man reading a book called <em>Medical English</em> asked us to join him for tea. He was mostly just trying to practice his English. He said he loves foreigners, and that his favorite period of life was the 80s and 90s - "We love Deng!".</p><p>My trip mate's friend's grandmother was incredibly gracious in hosting us for a dinner while we were in Emeishan. We saw a beautiful Uighur rug hanging on their wall. The grandfather pointed it out and explained, &#8220;We have amazing relationships with minorities here in China.&#8221; I don't think this was some theatrical attempt to contradict Western propaganda. It seemed like he genuinely didn&#8217;t know what is said about the treatment of Uighurs in the West. He was just trying to show his visitors something cool he had.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!UD_8!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8adb8d3e-2963-4f36-9455-fc1c75de5d3b_3384x2996.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!UD_8!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8adb8d3e-2963-4f36-9455-fc1c75de5d3b_3384x2996.jpeg 424w, https://substackcdn.com/image/fetch/$s_!UD_8!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8adb8d3e-2963-4f36-9455-fc1c75de5d3b_3384x2996.jpeg 848w, https://substackcdn.com/image/fetch/$s_!UD_8!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8adb8d3e-2963-4f36-9455-fc1c75de5d3b_3384x2996.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!UD_8!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8adb8d3e-2963-4f36-9455-fc1c75de5d3b_3384x2996.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!UD_8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8adb8d3e-2963-4f36-9455-fc1c75de5d3b_3384x2996.jpeg" width="3384" height="2996" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8adb8d3e-2963-4f36-9455-fc1c75de5d3b_3384x2996.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:2996,&quot;width&quot;:3384,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2168998,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!UD_8!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8adb8d3e-2963-4f36-9455-fc1c75de5d3b_3384x2996.jpeg 424w, https://substackcdn.com/image/fetch/$s_!UD_8!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8adb8d3e-2963-4f36-9455-fc1c75de5d3b_3384x2996.jpeg 848w, https://substackcdn.com/image/fetch/$s_!UD_8!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8adb8d3e-2963-4f36-9455-fc1c75de5d3b_3384x2996.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!UD_8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8adb8d3e-2963-4f36-9455-fc1c75de5d3b_3384x2996.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>We kept trying to ask them about their personal experiences over the last seven decades as China has grown and changed. but the grandfather kept responding with lengthy monologues about military history. Here is a representative exchange:</p><p>Me (who doesn&#8217;t speak Chinese) to my trip mate (who does): Ask them about what job they had when they first moved near Chengdu.</p><p>[The grandfather speaks in Chinese for 10 minutes. I get a bit impatient and whisper to my trip mate asking what he&#8217;s saying]</p><p>My trip mate: He's saying that the mountains around Chengdu make it a perfect base of retreat in case of an invasion, which Mao was worried about during the Sino Soviet split.</p><p>Me: What?! Why didn&#8217;t you ask him about the first job he had?</p><p>My trip mate: I did!!</p><p>Grandpas are the same everywhere.</p><p>People say that Xi is establishing a cult of personality. This may be true within the CCP cadres, but I saw no evidence of it in public. I don't think I saw a single picture of Xi anywhere - not on any billboards, screens, or walls. People didn't really bring Xi up in conversations. I saw some pictures of Mao, but mostly in museums (or in one case at a tea farm he apparently used to frequent). The hammer and sickle was also a rare sight, mostly displayed on government buildings.</p><p>There are indeed cameras everywhere. This is gonna sound super naive - but I genuinely don't understand why. There's no crime. I know you'll say it's to prevent protests. Which might make sense for major streets, but even random alleyway corners will have a couple of cameras. Are they really trying to prevent someone from fomenting insurrection between 2 garbage cans? Beijing in particular had police officers at attention at what seemed like every street corner.</p><p>One recent student I talked to said that they understand they don't have freedoms here, but they're willing to take the tradeoff in favor of safety - they don't want school shootings. I thought this was quite silly - not only because there's no reason political freedom ought to lead to school shootings, but mostly because school shootings are such a statistically marginal experience. But I realized this is exactly the way we treat any hints of public protests in China. Just as school shootings are featured heavily in the media but aren't actually something you're likely to personally encounter, so too with protests against the CCP. You are overwhelmingly unlikely to spontaneously encounter them.</p><p>People were quite willing to chat openly in public places about problems in the country and with the regime. Including people who seemed to have a lot to lose. Almost everyone I talked to would acknowledge the economy was bad, and many were willing to implicate the government's decisions. Some even casually brought up Tiananmen or the Cultural Revolution. One person was even willing to discuss the odds of regime change at a public restaurant - though he may been have an especially careless fellow.</p><p>To be clear, it's an authoritarian system, and I certainly would feel uncomfortable doing what I'm doing there, but it definitely isn't North Korea.</p><h2>Youngsters</h2><p>In a shopping mall in Chongqing, a couple of high schoolers came up to us in order to get selfies. It felt like the perfect opportunity to learn about young adult life in China. So I whipped out the Translate app in WeChat and proceeded, rather clumsy, to make small talk. I asked them what they did in their free time. They said they watched 2-3 hours of TikTok every day. I asked them what videos they'd watch. They said it's a whole bunch of "sexy girls". I laughed because I thought they were joking. I asked one of them to pull out his phone. He scrolled past the first 10 videos on his feed and they were indeed all just "sexy girls".</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Az_Q!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35dcd7fe-cf0a-4b88-9535-45e6713ff223_3024x4032.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Az_Q!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35dcd7fe-cf0a-4b88-9535-45e6713ff223_3024x4032.png 424w, https://substackcdn.com/image/fetch/$s_!Az_Q!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35dcd7fe-cf0a-4b88-9535-45e6713ff223_3024x4032.png 848w, https://substackcdn.com/image/fetch/$s_!Az_Q!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35dcd7fe-cf0a-4b88-9535-45e6713ff223_3024x4032.png 1272w, https://substackcdn.com/image/fetch/$s_!Az_Q!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35dcd7fe-cf0a-4b88-9535-45e6713ff223_3024x4032.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Az_Q!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35dcd7fe-cf0a-4b88-9535-45e6713ff223_3024x4032.png" width="1456" height="1941" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/35dcd7fe-cf0a-4b88-9535-45e6713ff223_3024x4032.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1941,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:21374535,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Az_Q!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35dcd7fe-cf0a-4b88-9535-45e6713ff223_3024x4032.png 424w, https://substackcdn.com/image/fetch/$s_!Az_Q!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35dcd7fe-cf0a-4b88-9535-45e6713ff223_3024x4032.png 848w, https://substackcdn.com/image/fetch/$s_!Az_Q!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35dcd7fe-cf0a-4b88-9535-45e6713ff223_3024x4032.png 1272w, https://substackcdn.com/image/fetch/$s_!Az_Q!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35dcd7fe-cf0a-4b88-9535-45e6713ff223_3024x4032.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">These guys love sexy girls</figcaption></figure></div><p>We chatted up quite a lot of young people on night life streets. I was struck by how many young people expressed feeling stressed or overwhelmed. We met a musician in Chengdu who was writing songs about youth anxiety. We chatted up some modeling school students - even they complained about the intense pressure they felt. We met a guy who had studied in Australia but returned to China during COVID. He explained that many of his friends with prestigious degrees are moving away from Shanghai and Beijing - Yes, the pay there can be twice as high as in second or third tier cities. But the competitiveness is insane. And in order to actually land the high skilled positions, they have to work truly insane hours (9-9-6 is not a myth). He said that many of his friends were opting for these less ambitious lower-paying careers in smaller cities, where the rent is lower and the pressure is manageable.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!wYz6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae058be0-5a01-4868-8ab5-c9c7baa20f3e_1352x1296.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!wYz6!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae058be0-5a01-4868-8ab5-c9c7baa20f3e_1352x1296.png 424w, https://substackcdn.com/image/fetch/$s_!wYz6!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae058be0-5a01-4868-8ab5-c9c7baa20f3e_1352x1296.png 848w, https://substackcdn.com/image/fetch/$s_!wYz6!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae058be0-5a01-4868-8ab5-c9c7baa20f3e_1352x1296.png 1272w, https://substackcdn.com/image/fetch/$s_!wYz6!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae058be0-5a01-4868-8ab5-c9c7baa20f3e_1352x1296.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!wYz6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae058be0-5a01-4868-8ab5-c9c7baa20f3e_1352x1296.png" width="1352" height="1296" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ae058be0-5a01-4868-8ab5-c9c7baa20f3e_1352x1296.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1296,&quot;width&quot;:1352,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2704281,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!wYz6!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae058be0-5a01-4868-8ab5-c9c7baa20f3e_1352x1296.png 424w, https://substackcdn.com/image/fetch/$s_!wYz6!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae058be0-5a01-4868-8ab5-c9c7baa20f3e_1352x1296.png 848w, https://substackcdn.com/image/fetch/$s_!wYz6!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae058be0-5a01-4868-8ab5-c9c7baa20f3e_1352x1296.png 1272w, https://substackcdn.com/image/fetch/$s_!wYz6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae058be0-5a01-4868-8ab5-c9c7baa20f3e_1352x1296.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Karaoke in Shanghai</figcaption></figure></div><p>Speaking of which - we really need to do a better job with Chinese students studying abroad in America. These students will likely end up in influential positions back home, yet colleges treat them basically like cash cows. They often arrive bombarded with propaganda about America, which is reinforced by the prevalent discourse at universities. And they find themselves isolated by language barriers and cultural differences. Giving these future leaders a genuinely positive experience in America might be the best thing we can do to improve US-China relations in the long run.</p><p>I'm still puzzled by how China can have both a demographic collapse and massive youth unemployment. You'd think with fewer young people being born, the ones who are around would be in high demand. One explanation I heard while there is that there are plenty of menial jobs available, but today's educated youth - who've gone through high school and college - just won't take the low-skilled positions their parents and grandparents did. Meanwhile, there's a real shortage of the high-skilled jobs that would actually match their education and aspirations. It's a mismatch between the jobs available and the jobs young people feel qualified for and willing to do.</p><p>I kept asking young people about the public intellectual landscape in China - who are their equivalents of Jordan Peterson, Joe Rogan, Lex Friedman, and Sam Harris? The sense I got is that this kind of popular intellectual ecosystem just doesn't exist there. Sure, there are viral Bilibili videos from professors talking about practical matters like how to manage your finances. But grand takes about what's happening in the world and what we should do about it? Not much going on.</p><p>I met a couple of EAs in China - they're exceptionally rare. When we asked them why it's so hard to spread EA or similar worldviews there, they said people just aren't into that organizing decisions using some ideological lens. They're much more concerned with practical, tangible matters. One talking point I heard again and again is that, "China is still a developing country." The implication seemed to be that China can't afford to pursue costly idealistic programs around climate change, social safety nets, or international aid - it needs to focus on basic economic development first.</p><h2>The Great Firewall</h2><p>By far the most inconvenient thing about visiting China is internet access. As a foreigner, basically all the websites you might find useful are behind the firewall. Even for websites that aren't blocked by the Great Firewall, I just didn't have anything even close to what you might call high-speed internet. This was true across my entire travel, including my burner SIM card from ChinaMobile or the WiFi at fancy hotels in Beijing or Shanghai. We released an episode of my podcast during this trip - because of the low bandwidth, my editor and I were just sending screenshots to each other throughout the night instead of hopping on a video call. The VPN situation is worse than I thought it would be . I happened to be lucky enough to download one of the VPNs that sometimes works before the trip. It was still pretty slow and unreliable. Astrill or Mullvad might actually result in a connection 75% of the time - often they would cause WeChat and Alipay to crash (which is rather inconvenient since you basically need those apps open all the time to navigate China). Very possible that I was fucking something up. But between these internet issues and the need to use a burner phone and laptop, I would be quite reluctant to use China as a remote work location.</p><h2>Tech &amp; AI</h2><p>I am super hesitant to say anything here, because I am extremely unconfident about what's actually happening. Treat this just as some tentative notes from a few conversations.</p><p>I've cut out a whole bunch of stuff about AI in this section because I'm not sure if my initial assessment is correct.  I hope to make podcast episodes over the next few months which provide a better researched account than I can supply here.</p><p>The biggest surprise from talking to Chinese VCs people at AI labs was how capital constrained they felt. Moonshot AI, one of China's leading AI labs, raised $1 billion at a $3 billion valuation. Meanwhile, just xAI's new cluster alone will cost $3-4 billion.</p><p>The tech ecosystem feels quite shell shocked from the 2021 crackdown. One VC half-jokingly asked if I could help him get his money out of China. If you keep your money in China, you're basically stuck choosing between terrible options. You can either accept a measly 2% yield from state banks, or throw it into China's perpetually struggling stock market. This helps explain why valuations for Chinese companies are chronically low - the exit opportunities just suck. Even if you build (or invest in) something great, there's no guarantee the company will be able to raise the next round. And even if you do raise again and succeed, the government might randomly cancel your IPO. And even if you somehow make it to the public markets, Chinese equities have been performing terribly anyways. It's a good reminder of how easy it is to completely wreck an innovation ecosystem that depends on risk-taking investors.</p><h2>Hearts and Minds</h2><p>In China, liberal pro-Western voices are often censored or shouted down. If I was the US President, and I wanted to win hearts and minds in China, here's what I'd do. In every single speech where I'm talking about China, I'd make a conspicuous effort to complement Chinese people, Chinese values, and Chinese culture. I'd talk about how my Chinese staffers are the smartest and most hardworking people I've ever worked with (which honestly is probably true). I'd talk about how much my daughter is obsessed with ancient Chinese dresses. I'd talk about how I'm learning Mandarin in my free time, and have a live "Aw shucks" conversation in Mandarin.</p><p>These clips would go viral on Bilibili and TikTok. And they'd probably stay up because it would just be a weird thing to censor. The CCP might even think that these displays of affection aggrandize them. But in reality, showing our admiration for Chinese people and their achievements (who genuinely are fucking killing it everywhere where they're not held down by communism), undermines the central narrative of the regime - that the West is hell bent on holding Chinese people back, that they have no respect or understanding of their culture, and that the CCP is a necessary bulwark against these imperialists.</p><p>On a totally unrelated note, so many people I met in China are in fact super talented and hard working. Someone connected me with the CEO of a company that manufactures life sciences equipment. He told me that before he started this company, he used to be a repairman for foreign imported machines, traveling 200 days a year with a screwdriver to wherever one had broken down. In 2013, he designed and started manufacturing his own machines, building more and more advanced designs, and now they've just built out a $60 million factory 1.5 hours from Shanghai. I asked him what the hardest part of ramping up production is - apparently constructing factories is just not an issue - tons of construction firms can make you a new facility quite reliably (along with the adjacent dorm room building for workers).</p><h2>What is travel good for?</h2><p>Noah Smith has a good <a href="https://www.noahpinion.blog/p/how-much-can-you-really-learn-about">blog post</a> about what one can't learn from travel - you're not going to learn about the risk of a war or the state of the AI race by gazing at skylines or chatting up taxi drivers. Of course you can learn about those things by talking to the princelings and researchers and CEOs. But if you have access to these higher ups, surely you can also get them on the Zoom call. And fwiw, this should you update you in favor of more Zoom calls, not less travel. During the trip I realized how much I could have already learned about China by meeting the listeners I have in China (who are unusually high quality, presumably because of the restrictions they have to get through to get to my content).</p><p>Incidentally, if you're in China (and especially if you work on anything related to AI) please email me at <a href="mailto:hello@dwarkeshpatel.com">hello@dwarkeshpatel.com</a>. I'm keen for people to reach out and suggest guests to chat about China. Especially keen to better understand the Chinese political system, and how decisions about mega projects, arms races, and technology investment will be made at the brink of AGI.</p><p>So what's the point of travel? What about the in person experience can't be replaced by books, travel vlogs, and Zoom calls? For me, it's something like, what becomes salient to you. I started asking questions about China I hadn't even thought to ask about America.</p><p>Another thing I noticed is more personal. Two weeks of being AFK, and of having the excuse of using a burner phone to put off messages, helped me clear the cache of thoughts about sponsorships, logistics, growth, hiring, and a bunch of other practical minutiae. My shower thoughts wandered away from upcoming negotiations and towards interesting rabbit holes.</p><p>It's a good reminder that what's lacking in life is not time. It's focus. If you're working on what matters, you can advance leaps and bounds in 8 hours. And if you're just clearing the slog, you can spend a lifetime staying in the same place.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!bulU!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fb66771-f2bb-4d91-899f-5490463cd90b_4032x3024.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!bulU!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fb66771-f2bb-4d91-899f-5490463cd90b_4032x3024.jpeg 424w, https://substackcdn.com/image/fetch/$s_!bulU!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fb66771-f2bb-4d91-899f-5490463cd90b_4032x3024.jpeg 848w, https://substackcdn.com/image/fetch/$s_!bulU!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fb66771-f2bb-4d91-899f-5490463cd90b_4032x3024.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!bulU!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fb66771-f2bb-4d91-899f-5490463cd90b_4032x3024.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!bulU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fb66771-f2bb-4d91-899f-5490463cd90b_4032x3024.jpeg" width="1456" height="1092" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2fb66771-f2bb-4d91-899f-5490463cd90b_4032x3024.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1092,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1991218,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!bulU!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fb66771-f2bb-4d91-899f-5490463cd90b_4032x3024.jpeg 424w, https://substackcdn.com/image/fetch/$s_!bulU!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fb66771-f2bb-4d91-899f-5490463cd90b_4032x3024.jpeg 848w, https://substackcdn.com/image/fetch/$s_!bulU!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fb66771-f2bb-4d91-899f-5490463cd90b_4032x3024.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!bulU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fb66771-f2bb-4d91-899f-5490463cd90b_4032x3024.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.dwarkesh.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.dwarkesh.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[Review of 2 books: China’s Economy by Arthur Kroeber & China Chapter in Trade Wars are Class Wars by Pettis & Klein ]]></title><description><![CDATA[Chapter on China in Trade Wars are Class Wars by Michael Pettis and Matthew Klein]]></description><link>https://www.dwarkesh.com/p/chinas-economy</link><guid isPermaLink="false">https://www.dwarkesh.com/p/chinas-economy</guid><dc:creator><![CDATA[Dwarkesh Patel]]></dc:creator><pubDate>Tue, 26 Nov 2024 18:17:16 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fad553126-18dc-40bc-a864-c1622faa3b06_1512x972.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>One thing I should clarify. I choose books to review which cover topics that I am especially unfamiliar of. I'm definitely not claiming any sort of expertise whatsoever on this topic. I'm just making tentative notes and judgments which I'll update over time as I learn more. </p><h2>Chapter on China in <em>Trade Wars are Class Wars</em> by Michael Pettis and Matthew Klein </h2><p>When China was booming, naive Westerners loved saying, "Our leaders think in 4 year terms - the Chinese government thinks in decades and centuries". In practice, authoritarian states (including China) are often far more short termist than market based democracies, because signals about the rational allocation of resources (like what people want to invest in, how they value different goods, where they want to live and work) are systematically repressed for political benefit.</p><p>Michael Pettis and Matthew Klein explain this incredibly well in their new book, <em><a href="https://www.amazon.com/Trade-Wars-Are-Class-International-ebook/dp/B087TJKJRQ">Trade Wars are Class Wars</a></em>. I learned more from simply their one chapter about China than I did from the the sum of every other book I read about the Chinese economy.</p><blockquote><p>In China, the GDP growth rate is an input into the system. It is set early in the year as the GDP growth target for that year and represents the amount of growth needed to accommodate social and political objectives, among which of course is the desire to keep unemployment low &#8230; The easiest way for officials to hit their targets is therefore to tell the state-run banks to lend to favored companies to invest in as much infrastructure, manufacturing, and real estate as necessary. Whether the investments are worthwhile is irrelevant. All that matters is that the quantity of spending generates enough reported GDP to meet the central government&#8217;s objectives.</p></blockquote><p>When I was in China, an investor (somewhat) jokingly asked if I could help him get his money out. He faced terrible options: put savings in state banks for a 2% yield, invest in China's perpetually struggling stock market, or... that's basically it, since capital controls trap money inside China. This creates a stealth tax on savers - you&#8217;re not allowed to start competitive bank offering real market interest rates from lending to productive private companies. Instead, state banks funnel money to local government projects and state enterprises. They might build those empty cities and bridges to nowhere "efficiently," but without market pressure to make actually useful investments, the capital often gets wasted.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.dwarkesh.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe for more posts</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>This financial repression<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> leads not only to massive malinvestment (which the economy is now paying for) but also redistribution from households to state favored companies.</p><p>I remain confused about how this fits into the bigger point that Pettis makes, which is that the Chinese government is causing distortions which lead to too much saving and not enough consumption. Isn't this financial repression an implicit tax on savers?  Maybe the government is doing a bunch of other things which more than offset this tax on savers? Adam Posen has a more compelling <a href="https://www.foreignaffairs.com/responses/who-killed-chinese-economy">explanation</a> - this glut of savings is not the result of particular macroeconomic policies. Rather it&#8217;s the public&#8217;s natural hunker-down response to the government&#8217;s capricious handling of Zero-Covid.</p><p>Heilman&#8217;s <em><a href="https://www.amazon.com/Chinas-Political-System-Sebastian-Heilmann/dp/1442277343">China&#8217;s Political System</a></em> has a very useful chart illustrating explicit taxes - as you can see, they&#8217;re quite low, and mostly dominated by value added tax and corporate taxes, with personal income taxes being negligible. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!49nl!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4988576-5126-496d-860f-9becdb89e87d_1846x2228.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!49nl!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4988576-5126-496d-860f-9becdb89e87d_1846x2228.png 424w, https://substackcdn.com/image/fetch/$s_!49nl!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4988576-5126-496d-860f-9becdb89e87d_1846x2228.png 848w, https://substackcdn.com/image/fetch/$s_!49nl!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4988576-5126-496d-860f-9becdb89e87d_1846x2228.png 1272w, https://substackcdn.com/image/fetch/$s_!49nl!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4988576-5126-496d-860f-9becdb89e87d_1846x2228.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!49nl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4988576-5126-496d-860f-9becdb89e87d_1846x2228.png" width="1456" height="1757" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e4988576-5126-496d-860f-9becdb89e87d_1846x2228.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1757,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2644404,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!49nl!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4988576-5126-496d-860f-9becdb89e87d_1846x2228.png 424w, https://substackcdn.com/image/fetch/$s_!49nl!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4988576-5126-496d-860f-9becdb89e87d_1846x2228.png 848w, https://substackcdn.com/image/fetch/$s_!49nl!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4988576-5126-496d-860f-9becdb89e87d_1846x2228.png 1272w, https://substackcdn.com/image/fetch/$s_!49nl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4988576-5126-496d-860f-9becdb89e87d_1846x2228.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>But I think these implicit taxes (this penalty on savers, or the land sales on expropriated property<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>) are indeed taxes. So I asked Claude to chart them. Claude came up with these numbers on its own - so take them with a grain of salt.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!69qa!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7598eb6b-dbee-4e80-8aba-135e0118d310_1062x1082.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!69qa!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7598eb6b-dbee-4e80-8aba-135e0118d310_1062x1082.png 424w, https://substackcdn.com/image/fetch/$s_!69qa!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7598eb6b-dbee-4e80-8aba-135e0118d310_1062x1082.png 848w, https://substackcdn.com/image/fetch/$s_!69qa!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7598eb6b-dbee-4e80-8aba-135e0118d310_1062x1082.png 1272w, https://substackcdn.com/image/fetch/$s_!69qa!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7598eb6b-dbee-4e80-8aba-135e0118d310_1062x1082.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!69qa!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7598eb6b-dbee-4e80-8aba-135e0118d310_1062x1082.png" width="1062" height="1082" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7598eb6b-dbee-4e80-8aba-135e0118d310_1062x1082.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1082,&quot;width&quot;:1062,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:160534,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!69qa!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7598eb6b-dbee-4e80-8aba-135e0118d310_1062x1082.png 424w, https://substackcdn.com/image/fetch/$s_!69qa!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7598eb6b-dbee-4e80-8aba-135e0118d310_1062x1082.png 848w, https://substackcdn.com/image/fetch/$s_!69qa!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7598eb6b-dbee-4e80-8aba-135e0118d310_1062x1082.png 1272w, https://substackcdn.com/image/fetch/$s_!69qa!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7598eb6b-dbee-4e80-8aba-135e0118d310_1062x1082.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Up until the 1990s, this state led investment paradigm was fine:</p><blockquote><p>At least until the mid-1990s, this incentive system was not a problem for China because the shortage of infrastructure and manufacturing capacity was so large. The only important constraint on productive investment was the pace at which savings could grow. Almost any investment increased productivity by far more than the cost of the project.</p></blockquote><p>But afterwards this dynamic began to change. In response to the 2008 financial crisis, the Chinese government doubled down on its investment led growth model, launching a massive 568 billion dollar stimulus. The subsequent malinvestment has been so pervasive that total factor productivity in China, which is already significantly lower than America's, has actually decreased.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!spWj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fad553126-18dc-40bc-a864-c1622faa3b06_1512x972.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!spWj!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fad553126-18dc-40bc-a864-c1622faa3b06_1512x972.png 424w, https://substackcdn.com/image/fetch/$s_!spWj!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fad553126-18dc-40bc-a864-c1622faa3b06_1512x972.png 848w, https://substackcdn.com/image/fetch/$s_!spWj!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fad553126-18dc-40bc-a864-c1622faa3b06_1512x972.png 1272w, https://substackcdn.com/image/fetch/$s_!spWj!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fad553126-18dc-40bc-a864-c1622faa3b06_1512x972.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!spWj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fad553126-18dc-40bc-a864-c1622faa3b06_1512x972.png" width="1456" height="936" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ad553126-18dc-40bc-a864-c1622faa3b06_1512x972.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:936,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:280002,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!spWj!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fad553126-18dc-40bc-a864-c1622faa3b06_1512x972.png 424w, https://substackcdn.com/image/fetch/$s_!spWj!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fad553126-18dc-40bc-a864-c1622faa3b06_1512x972.png 848w, https://substackcdn.com/image/fetch/$s_!spWj!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fad553126-18dc-40bc-a864-c1622faa3b06_1512x972.png 1272w, https://substackcdn.com/image/fetch/$s_!spWj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fad553126-18dc-40bc-a864-c1622faa3b06_1512x972.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Paul Krugman wrote an excellent essay in 1997 called <em><a href="https://www.brmandel.com/uploads/3/2/4/5/3245755/myth_of_asias-miracle.pdf">The Myth of The Asian Miracle</a></em>,<em> </em>which explains why an increase in productivity, as opposed to a buildup of inputs alone, is crucial for long run economic growth. He explains how the initially impressive period of Soviet growth was inherently limited, and why the same dynamic would constrain the growth of China and Japan:</p><blockquote><p>sustained growth in a nation&#8217;s per capita income can only occur if there is a rise in output per unit of input.</p><p>Mere increases in inputs, without an increase in the efficiency with which those inputs are used&#8212;investing in more machinery and infrastructure&#8212;must run into diminishing returns; input-driven growth is inevitably limited&#8230;</p><p>Even a modest slowing in China&#8217;s growth will change the geopolitical outlook substantially. The World Bank estimates that the Chinese economy is currently about 40 percent as large as that of the United States. Suppose that the U.S. economy continues to grow at 2.5 percent each year. If China can continue to grow at 10 percent annually, by the year 2010 its economy will be a third larger than ours. But if Chinese growth is only a more realistic 7 percent, its GDP will be only 82 percent of that of the United States. There will still be a substantial shift of the world&#8217;s economic center of gravity, but it will be far less drastic than many people now imagine.</p></blockquote><p>Btw, people give Krugman tremendous amounts of shit for underestimating the economic impact of the Internet. But calling China&#8217;s stagnation 3 decades early, during a period in which it was pretty contrarian, is actually super impressive. </p><p>All these loans from state owned banks to often unproductive investment amount to a stupendous total debt load for what is still a developing country. The following chart was made by Claude, including the numbers, so again, grain of salt.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!E1iJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd90d455c-00a6-4099-b340-b7b35f82e089_1098x812.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!E1iJ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd90d455c-00a6-4099-b340-b7b35f82e089_1098x812.png 424w, https://substackcdn.com/image/fetch/$s_!E1iJ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd90d455c-00a6-4099-b340-b7b35f82e089_1098x812.png 848w, https://substackcdn.com/image/fetch/$s_!E1iJ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd90d455c-00a6-4099-b340-b7b35f82e089_1098x812.png 1272w, https://substackcdn.com/image/fetch/$s_!E1iJ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd90d455c-00a6-4099-b340-b7b35f82e089_1098x812.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!E1iJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd90d455c-00a6-4099-b340-b7b35f82e089_1098x812.png" width="1098" height="812" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d90d455c-00a6-4099-b340-b7b35f82e089_1098x812.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:812,&quot;width&quot;:1098,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:93047,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!E1iJ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd90d455c-00a6-4099-b340-b7b35f82e089_1098x812.png 424w, https://substackcdn.com/image/fetch/$s_!E1iJ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd90d455c-00a6-4099-b340-b7b35f82e089_1098x812.png 848w, https://substackcdn.com/image/fetch/$s_!E1iJ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd90d455c-00a6-4099-b340-b7b35f82e089_1098x812.png 1272w, https://substackcdn.com/image/fetch/$s_!E1iJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd90d455c-00a6-4099-b340-b7b35f82e089_1098x812.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>A future Chinese reformer would actually face a tougher challenge than Deng did in 1978. This seems like a strange thing to say - surely being the second largest economy in the world puts you in a better position to continue further growth. But Japan and the Soviet Union showed that when a mature economy crashes from a debt bubble, recovery is brutal. Deng, in contrast, got to start from scratch with several advantages: minimal existing debt made borrowing cheap; barely any existing infrastructure meant new investments were super productive and maintenance costs were low; plus China cashed in on history&#8217;s largest demographic dividend with tons of young workers and falling birth rates.</p><p>Now all these advantages have flipped into disadvantages, making any future reforms much less likely to deliver any quick wins.</p><p>This isn&#8217;t the most important consequence, but aesthetically this is a tragedy. The 3 decade construction frenzy which ended just a couple years ago put up endless shoddy looking skyscrapers everywhere. If this is the end of major construction for a while, then they&#8217;ve built a country that is uglier that it might have otherwise been.</p><p>Pettis and Klein offer a compelling take on China's Belt and Road Initiative: it's not primarily about building political alliances - it's about exporting China's domestic investment model abroad. Chinese companies get state subsidies to build infrastructure projects overseas, just like they do at home. But there's a key difference: when these projects turn out to be unproductive (think empty ports or underused railways), it's the host countries that end up stuck with the maintenance costs and depreciation, not Beijing. It's essentially a pressure release valve for China's overcapacity in construction and heavy industry.</p><p>The 2008 crisis seemed to vindicate the worldview that the free market model was inherently flawed. Instead of pressing hard on reforms, which would have been much more palatable before this huge buildup of debt and malinvestment, the Chinese government doubled down on state led investment and infrastructure.</p><p>In 1961, Nobel laureate Paul Samuelson confidently predicted in his economics textbook that the Soviet economy would overtake America's by the 1980s or '90s. During Japan's rise, Western experts couldn't stop gushing about how MITI's central planning was the secret sauce behind their economic miracle. More recently, books like "How Asia Works" praised capital controls and industrial policy - ideas that look a lot less brilliant now that we're watching their consequences play out in China.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Q-QM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cdf056f-b5cb-4911-a6e8-9c449d9fa2a5_389x613.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Q-QM!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cdf056f-b5cb-4911-a6e8-9c449d9fa2a5_389x613.png 424w, https://substackcdn.com/image/fetch/$s_!Q-QM!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cdf056f-b5cb-4911-a6e8-9c449d9fa2a5_389x613.png 848w, https://substackcdn.com/image/fetch/$s_!Q-QM!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cdf056f-b5cb-4911-a6e8-9c449d9fa2a5_389x613.png 1272w, https://substackcdn.com/image/fetch/$s_!Q-QM!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cdf056f-b5cb-4911-a6e8-9c449d9fa2a5_389x613.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Q-QM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cdf056f-b5cb-4911-a6e8-9c449d9fa2a5_389x613.png" width="389" height="613" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9cdf056f-b5cb-4911-a6e8-9c449d9fa2a5_389x613.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:613,&quot;width&quot;:389,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;samuelson&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="samuelson" title="samuelson" srcset="https://substackcdn.com/image/fetch/$s_!Q-QM!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cdf056f-b5cb-4911-a6e8-9c449d9fa2a5_389x613.png 424w, https://substackcdn.com/image/fetch/$s_!Q-QM!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cdf056f-b5cb-4911-a6e8-9c449d9fa2a5_389x613.png 848w, https://substackcdn.com/image/fetch/$s_!Q-QM!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cdf056f-b5cb-4911-a6e8-9c449d9fa2a5_389x613.png 1272w, https://substackcdn.com/image/fetch/$s_!Q-QM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cdf056f-b5cb-4911-a6e8-9c449d9fa2a5_389x613.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">&#8220;In the 1961 edition of his famous textbook of economic principles, Paul Samuelson wrote that GNP in the Soviet Union was about half that in the United States but the Soviet Union was growing faster.&nbsp; As a result, one could comfortably forecast that Soviet GNP would exceed that of the United States by as early as 1984 or perhaps by as late as 1997 and in any event Soviet GNP would greatly catch-up to U.S. GNP.&#8221; - <a href="https://marginalrevolution.com/marginalrevolution/2010/01/soviet-growth-american-textbooks.html">Alex Tabarrok on MR</a></figcaption></figure></div><p>Maybe Bryan Caplan had it right when he <a href="https://www.econlib.org/libertys-crisis-crisis/">said</a> that libertarians do indeed have &#8220;timeless ideas for a better world&#8221;. Maybe we should stop getting excited every time we think we've found an exception to market principles, which again eventually joins the pile of disappointments.</p><h2><em>China&#8217;s Economy: <strong>What Everyone Needs to Know</strong></em> by Arthur R. Kroeber</h2><ul><li><p>Advocates of an authoritarian model (who are particularly inspired by China) don't take seriously enough fragile good government is in these societies. You go from Mao (on a death toll basis, literally the worst government in the history of human kind - 50m+ dead) to Deng (inspired leadership) to Jiang Zemin (who was kept on the path of reform by the power behind the throne Deng) to Hu Jintao (2008+ property bubble, malinvestment, etc that may leave China stuck as middle income country) to Xi (zero Covid, lack of growth, centralization of power). A system of government which requires a Deng type figure to wrestle against all odds to provide great leadership for 20-30 years, and then returns to the same-old same-old is not optimal!</p></li><li><p>Is there a single example of a country which experienced rapid industrialization and fast growth which was a full democracy at the time? Japan during Meiji was not, neither was SK/Taiwan/China during periods of highest growth. UK and US were very partial democracies. Japanese and German growth after WW2 was very fast, but some of it was rebuilding what they already had.</p><ul><li><p>Deng is counterfactually crucial even after 1992. He basically told Jiang Zemin in his Southern tour that if he didn't keep up reforms, he'd replace him.</p></li></ul></li><li><p>Why is the example of SK/Taiwan/Japan not a valid argument for the idea that political reform follows economic reform? In those cases, political reform was the result of strong US pressure (or in the case of Japan, imposition). It's not clear that it would have happened by default.</p></li><li><p>China's cities have surprisingly good infrastructure for a middle-income country, but there's a specific reason why: it's all about how they finance it. In China, the government owns all land - when you "buy" property, you're really just getting a long-term lease. This lets provincial governments use a clever (but dangerous) funding trick: they borrow money against land they control, justifying high valuations by promising future infrastructure improvements will drive up prices. For years, this worked beautifully because property values kept soaring.</p><p>This system went into overdrive after 2008, when Beijing's stimulus plan essentially told local governments to go wild with infrastructure spending. They did exactly that, borrowing heavily against their land to fund endless construction projects. The result? Local governments are now sitting on a mountain of debt, as shown in the chart above. </p></li><li><p>People often point to China's explosive growth from 1980-2020 (averaging around 10% annually) as a model for potential AI takeoff. But how good is this analogy? Some argue this wasn't just about throwing more resources at the problem - until 2008, roughly three-quarters of China's growth came from Total Factor Productivity (TFP) increases, not just adding more labor or capital. After 2008, this shifted as government malinvestment took over.</p><p>But here's the thing: high TFP growth doesn't necessarily mean genuine innovation. TFP is just what's left over after accounting for labor and capital inputs - so even simple changes like "Mao dying and letting markets work" show up as TFP growth. Are we really expecting AIs to generate truly new ideas at the same rate that developing economies can copy existing ones from more advanced countries? That's a much bigger ask.</p></li><li><p>Tiananmen taught the Chinese Communist Party a crucial lesson: urban unrest is the biggest threat to regime survival. But there's an often-overlooked economic context to those protests - inflation was surging in the late 1980s. This wasn't just a political crisis, it was an economic one.</p><p>The inflation happened because of Deng's early reforms: they partially liberalized prices while still maintaining the dual-track system (where some prices were market-based while others remained state-controlled). This created opportunities for arbitrage between the two systems and, combined with rapid credit expansion and wage increases, sent prices soaring. By 1988-89, urban residents were watching their savings and purchasing power evaporate.</p><p>So when party leaders looked at Tiananmen's aftermath, they drew two conclusions: first, keep urban residents happy at all costs; second, maintain strict control over prices and financial stability. It's a classic case of economic hardship driving political upheaval.</p></li><li><p>If farmers in China had received the true market value of their land when expropriated by city governments, and were able to receive normal market returns on that investment, their wealth would be 8% of GDP higher. </p></li><li><p>Land reform is important for catchup growth because it increases agricultural yield, which allows for exports, which gives you foreign exchange, which you can use for machinery.</p></li><li><p>Export discipline is important for catch up growth because it forces firms to win by global standards (instead of political influence over domestic markets).</p></li><li><p>Why did China have to rely on FDI but not South Korea/Taiwan/Japan (who built up domestic champions instead)? Because SK, Japan, Taiwan was part of the US alliance, which gave them access to American technical know-how and market access. China wasn't part of this alliance, so had to rely on FDI.</p></li><li><p>The book nicely explained why value-added tax is more easily enforceable than other kinds of taxes. The merchant who sells the final product has a strong incentive to make sure his suppliers are reporting their sales accurately. Because the taxes on any difference between what the suppliers charge and what the final consumer pays has to be paid by the merchant. This dynamic plays out recursively with all the intermediate suppliers. </p></li><li><p>I realize that I didn't mention Hukou anywhere. I'd be curious if somebody has done an economic analysis of what percentage of GDP this system is costing China. Based on the naive <a href="https://pubs.aeaweb.org/doi/pdfplus/10.1257/jep.32.1.3">extrapolations</a> of the cost of rent being too high in American cities (which is much less intense than a rural/urban government registry), getting rid of Hukou might well increase China&#8217;s GDP over 50%.</p></li></ul><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.dwarkesh.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.dwarkesh.com/subscribe?"><span>Subscribe now</span></a></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Which was even worse when the RMB was devalued (after all, a devalued peg is essentially a subsidy for exporters funded by a tax on importers).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>A Georgist analysis of China's property market would be quite interesting to read. </p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[Livestream Sarah Paine tonight at 7 PM Pacific on Japan at War]]></title><description><![CDATA[Just download the app and subscribe to me. At 7 PM Pacific you'll see the live video front & center in the app.]]></description><link>https://www.dwarkesh.com/p/japan-at-war-announcement</link><guid isPermaLink="false">https://www.dwarkesh.com/p/japan-at-war-announcement</guid><dc:creator><![CDATA[Dwarkesh Patel]]></dc:creator><pubDate>Wed, 23 Oct 2024 21:04:56 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!5u1l!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3e33418-4770-4398-8e33-884399df4295_1358x938.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The first lecture and live interview with Professor Sarah Paine is tonight!</p><p>This one on Japan during WW2. If you can't make it in person, you can livestream it via the <a href="https://substack.com/app">Substack app</a>. </p><p>Just <a href="https://substack.com/app">download the app</a> and subscribe to me. Around 7 PM Pacific you'll see the live video front &amp; center on the homepage in the app.</p><p>Very excited!</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!5u1l!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3e33418-4770-4398-8e33-884399df4295_1358x938.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!5u1l!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3e33418-4770-4398-8e33-884399df4295_1358x938.png 424w, https://substackcdn.com/image/fetch/$s_!5u1l!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3e33418-4770-4398-8e33-884399df4295_1358x938.png 848w, https://substackcdn.com/image/fetch/$s_!5u1l!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3e33418-4770-4398-8e33-884399df4295_1358x938.png 1272w, https://substackcdn.com/image/fetch/$s_!5u1l!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3e33418-4770-4398-8e33-884399df4295_1358x938.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!5u1l!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3e33418-4770-4398-8e33-884399df4295_1358x938.png" width="1358" height="938" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d3e33418-4770-4398-8e33-884399df4295_1358x938.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:938,&quot;width&quot;:1358,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1321387,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!5u1l!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3e33418-4770-4398-8e33-884399df4295_1358x938.png 424w, https://substackcdn.com/image/fetch/$s_!5u1l!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3e33418-4770-4398-8e33-884399df4295_1358x938.png 848w, https://substackcdn.com/image/fetch/$s_!5u1l!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3e33418-4770-4398-8e33-884399df4295_1358x938.png 1272w, https://substackcdn.com/image/fetch/$s_!5u1l!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3e33418-4770-4398-8e33-884399df4295_1358x938.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div>]]></content:encoded></item></channel></rss>