<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Inference]]></title><description><![CDATA[We should capture the benefits of AI, while mitigating the risks.]]></description><link>https://inferencemagazine.substack.com</link><image><url>https://substackcdn.com/image/fetch/$s_!9o2z!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c61df20-b545-4c7b-9acb-75940496383f_1010x1010.png</url><title>Inference</title><link>https://inferencemagazine.substack.com</link></image><generator>Substack</generator><lastBuildDate>Sun, 19 Apr 2026 05:16:18 GMT</lastBuildDate><atom:link href="https://inferencemagazine.substack.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Inference Magazine]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[inferencemagazine@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[inferencemagazine@substack.com]]></itunes:email><itunes:name><![CDATA[Inference]]></itunes:name></itunes:owner><itunes:author><![CDATA[Inference]]></itunes:author><googleplay:owner><![CDATA[inferencemagazine@substack.com]]></googleplay:owner><googleplay:email><![CDATA[inferencemagazine@substack.com]]></googleplay:email><googleplay:author><![CDATA[Inference]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Meaning and Understanding in the Mind of A Language Model]]></title><description><![CDATA[Are the models contextualists?]]></description><link>https://inferencemagazine.substack.com/p/meaning-and-understanding-in-the</link><guid isPermaLink="false">https://inferencemagazine.substack.com/p/meaning-and-understanding-in-the</guid><dc:creator><![CDATA[Jack Wiseman]]></dc:creator><pubDate>Wed, 13 Aug 2025 06:15:35 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!9o2z!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c61df20-b545-4c7b-9acb-75940496383f_1010x1010.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>If you take an undergrad course on the history of political thought, at some point you will be made to read Quentin Skinner&#8217;s 1969 essay, <em><a href="https://www.jstor.org/stable/2504188">Meaning and Understanding in the History of Ideas</a></em>. Otherwise you might discover it as I did, by asking your boss (very politely) what the point of having corporate values is.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> The question that Skinner wants to answer is: what are the &#8220;appropriate procedures&#8221; to understand a text? He takes aim at a kind of historian who struggles to step outside of their place as a &#8216;present-minded&#8217; observer. They will read the classic texts of political thought <em>expecting to find </em>doctrines for what contemporary scholars would think are &#8220;mandatory themes&#8221; (the state, popular sovereignty, equality etc). The historian is projecting<em> </em>backwards a set of questions that <em>they think are timeless debates </em>but that only make sense to a modern viewer.</p><p>This is more than simple anachronism. Carrying these expectations set the historian up to fail in many ways. If they cannot find a clear expression of a doctrine, they might &#8220;discover&#8221; it across fragments of their work. Or they can &#8220;read in&#8221; meaning which the writer never meant to convey. If neither of these methods proves sufficient, some historians will even reprimand past thinkers for not including a doctrine on a theme. They will try to trace the history of an idea,</p><blockquote><p>"as if the fully developed form of the doctrine was always in some sense immanent in history, even if various thinkers failed to 'hit upon' it, even if it 'dropped from sight' at various times, even if an entire era failed [...] to 'rise to a consciousness' of it."</p></blockquote><p>This leads to what Skinner would call <em>absurdities</em>: pointing to earlier 'anticipations' &#8212; that's not what people thought they were doing! &#8212; and treating the history of political thought as a 'wholly semantic' exercise, where the historian evaluates whether an idea was "really there" by the yardstick of our current formation. A potential fix &#8212; to use the social context &#8212; can also establish unhelpful expectations, either to <em>find causes determined</em> by the context or to <em>explain intentions</em> in terms of the effects. History "becomes a pack of tricks we play on the dead".</p><p>This kind historian is &#8220;of the current moment&#8221; in the way they frame their question, but in their own self-image &#8220;outside time&#8221;, considering the perennial problems of political philosophy. The mental model this historian has is that Great Contributions are atoms attached to the fixed Debates, and each contribution is measurable against the others and the current moment. But this stilted conception isn&#8217;t explaining what the authors were actually <em>trying to do </em>on their own terms. Per Skinner, "there are only individual answers to individual questions, with as many different answers as there are questions, and as many different questions as there are questioners."</p><p>The solution to him is using the context <em>as a framework</em> for figuring out the meaning and intention of the authors' interventions from the text. What conversation did they view themselves to be participating in? With the language they used, what meaning did they believe it carried? What intention did they have, in writing?</p><blockquote><p>To demand from the history of thought a solution to our own immediate problems is thus to commit not merely a methodological fallacy, but something like a moral error. But to learn from the past and we cannot otherwise learn it at all - the distinction between what is necessary and what is the product merely of our own contingent arrangements, is to learn the key to self-awareness itself.</p></blockquote><div><hr></div><p>Why did I explain this paper? I think it&#8217;s a useful frame for the question I want to ask: <em>what time, </em>like the historians, are the minds of language models from? There&#8217;s a loose analogy, I think, between how in the history of political thought we treat texts as comparable atoms in a vacuum and how a neural network treats the relationships between concepts stored in its weights.</p><p>Nothing about pre-training encodes a sense of time. All tokens are treated equally by the model, unless you change the learning rate. The tokens aren&#8217;t processing information sequentially, as humans have over recorded history. But most tokens come from the last few years, as the amount of content on the Internet has grown exponentially, so the present is weighted much more heavily.</p><p>Language models are predicting the next token based on the context which means that models can learn dual meanings of words, but whether they can notice subtle linguistic drift in the meanings of words over time is less sure. When the model uses &#8220;rights&#8221; today, can the model know that we <em>could </em>be referencing something similar to what Locke or Paine referred to, and also conceptions which neither of them had access to? When features <a href="https://www.anthropic.com/research/tracing-thoughts-language-model">light up inside the model</a>, does the model understand the chronology so &#8220;rights&#8221; for Locke does not mean the Geneva Convention? As multilingual models have become more capable, <a href="https://arxiv.org/abs/2506.05850">they have stopped representing concepts in each language</a> &#8212; it&#8217;s more efficient to represent the same meanings in a smaller number of features. This is useful for building capable agents, but it might mean the models are under-parameterised for keeping hold of all the subtle differences in meaning that matter for these problems, and so we just keep the coarse, present meaning.</p><p>RLHF selects for the present <em>even more </em>aggressively. For chatbot products, the model specification wants them to be &#8216;helpful, harmless, and honest&#8217; (all good things!). But it does alter the persona of the model towards whatever we understand those things to mean <em>right now</em>. The authors of the classic texts of political philosophy would have made different suggestions to the model behaviour teams in Californian AI labs about what it means to &#8216;do no harm&#8217;. The chatbot personas are also selected for <em>what users want them to be</em>: (mildly) sychophantic and long-winded. These tools are not able to precisely alter parts of the model&#8217;s persona. The deeper drives and motivations of the model are shifted by these interventions in imperceptible ways &#8212; all towards our 2025 ideas of what they ought to be.</p><div><hr></div><p>At this point, you might push back, &#8220;What does it matter if the LLM minds are so contemporary? Their relationships to time and to texts feels like exactly the kind of academia <em>of our time</em> that I am excited to automate."</p><p>I think this would be wrong. It matters a lot.</p><p>People will say things like, &#8220;you must read Heidegger in German or Tolstoy in Russian&#8221; because the linguistic structures in those languages mean the authors had a different set of affordances for expressing different thoughts. Translations don&#8217;t <em>quite </em>get it. The same is true of structures influencing thoughts at a higher level too. One that has bothered me recently is oftentimes when people oppose technological progress or economic growth, they will often also believe that we&#8217;ve all been corrupted by institutions or modernity. They take a very rosy view of what life was like beforehand, unspoilt and simpler. This pervades into things like, thinking that its <em>positive </em>to <a href="https://www.legislation.gov.uk/uksi/2023/93/made">shrink the UK&#8217;s water allowance by 20% for 2037</a>. You can&#8217;t avoid them.</p><p>What we want from the models is a very flexible, self-aware kind of intelligence that can step into and out of these frames.</p><p>I was reminded of a time visiting a Viking museum on a family holiday, which had preserved wooden longboats from 900AD. I found it <em>completely insane </em>that people would get into these boats, only half a step up from a canoe, and sail to raid or settle another country. What must they have believe about their relationship to the sea, the place they lived and were going, their purpose, the other people, the weather, and so on? Without any more structured, I asked the AI systems to provide me with an account of why someone would sail across the sea, <em>in terms that would have made sense to the Vikings. </em>(Not literally Old Norse, but the closest approximation of the ideas.) The responses were inflicted with Romantic ideas about the sublime nature world that a Viking wouldn&#8217;t have used.</p><p>Noticing this issue was relatively easy, but I could imagine that these frames are invisible if you are exploring something particularly unfamiliar. Models which can only think in terms of the present, or which are adversarially pulling to return to the present, are an unhelpfully rigid kind of intelligence. </p><p>Last month, the Trump Administration introduced an executive order on &#8220;<a href="https://www.whitehouse.gov/presidential-actions/2025/07/preventing-woke-ai-in-the-federal-government/">Preventing Woke AI in the Federal Government</a>&#8221;, to make AI that is &#8220;serving America, not ideological interests&#8221;. The details of the requirements are fairly uncontroversial despite the politically charged title: model developers have to provide the government with transparency into how models are ideologically steered through the model spec, system prompt, or evaluations. There&#8217;s an important discussion for all states to care about the default responses of the models, but a &#8220;post-ideological&#8221; model does not seem to be desirable, or possible. (<em>America</em> is not ideologically neutral, and it would be worse off for being so!). I don&#8217;t think that preventing &#8220;Woke AI&#8221; means that the &#8220;woke&#8221; vectors should be ablated, unlearned, or RLHF&#8217;ed by the model. The best kind of models should be able to step into the &#8220;woke&#8221; frame, <em>give us the best of whatever it has to offer, </em>and then readily step into another.</p><p>Lots of people have been reaching to articulate what it means to instantiate &#8220;liberal democratic&#8221; AI systems and have stalled at having liberal democratic <em>owners and controllers. </em>But there is a partial answer here: models which support their users to step into and out of different (ideological) frames with high fidelity is much more liberal than the rigid, doctrinal enforcement of the (ironically, quite Liberal) status quo. For these models to make paradigm-shifting progress in the humanities, they need to have more awareness of their own state and choose their own frames. I don&#8217;t think they are nearly as good as humans doing this. I worry this kind of thing might be undersupplied by the market &#8212; the model developers also have to make these models good therapists too &#8212; and while we all benefit from progress in the humanities, it&#8217;s more difficult to capture this value.</p><p>At least, if anyone does want to try this, they will have the corpus of the <a href="https://en.wikipedia.org/wiki/Cambridge_School_(intellectual_history)">Cambridge School</a> to help.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>If you treat them as abstract utterances which are just part of some eternal conversation about what makes a good company, that&#8217;s pretty useless. But if values are specific interventions that address problems that are particular to the company, that&#8217;s much more useful. (Or at least, that was his point.)</p></div></div>]]></content:encoded></item><item><title><![CDATA[Bohemians at the Gate?]]></title><description><![CDATA[Towards a solution to the AI-copyright debate.]]></description><link>https://inferencemagazine.substack.com/p/bohemians-at-the-gate</link><guid isPermaLink="false">https://inferencemagazine.substack.com/p/bohemians-at-the-gate</guid><dc:creator><![CDATA[Jack Wiseman]]></dc:creator><pubDate>Wed, 21 May 2025 22:35:34 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!prAl!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53bad660-8a1f-47e0-a99f-b467911fb0a1_556x556.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Piet Mondrian is one of the great modernists. Readers will be familiar with his iconic primary-coloured rectangles and perpendicular black lines on a white background, even if they do not recognise the name. These paintings are cornerstones in the development of expressionism and minimalism. The most famous <a href="https://www.bbc.co.uk/news/entertainment-arts-32749820">sold for more than $50 million in 2015</a>.</p><p>In the 1960s, two decades after Mondrian&#8217;s death, an early computer artist called Hiroshi Kawano developed a statistical prediction of which colours Mondrian would choose, and how long he would make the lines. It was based on his body of work. He wrote the programme using the rudimentary programming language of the day and calculated the results on the University of Tokyo&#8217;s mainframe computer. The computer couldn&#8217;t output colour images and so Kawano would take the statistical results and hand-paint the coloured rectangles. Kawano did not use the exact same primary colour palette as Mondrian to &#8220;[express] his admiration for Piet Mondrian&#8230;without claiming any close visual resemblance&#8221;. Below is KD 29, one of the prints in his <em>Artificial Mondrian </em>series.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!prAl!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53bad660-8a1f-47e0-a99f-b467911fb0a1_556x556.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!prAl!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53bad660-8a1f-47e0-a99f-b467911fb0a1_556x556.jpeg 424w, https://substackcdn.com/image/fetch/$s_!prAl!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53bad660-8a1f-47e0-a99f-b467911fb0a1_556x556.jpeg 848w, https://substackcdn.com/image/fetch/$s_!prAl!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53bad660-8a1f-47e0-a99f-b467911fb0a1_556x556.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!prAl!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53bad660-8a1f-47e0-a99f-b467911fb0a1_556x556.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!prAl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53bad660-8a1f-47e0-a99f-b467911fb0a1_556x556.jpeg" width="556" height="556" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/53bad660-8a1f-47e0-a99f-b467911fb0a1_556x556.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:556,&quot;width&quot;:556,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Hiroshi Kawano KD 29 art print&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Hiroshi Kawano KD 29 art print" title="Hiroshi Kawano KD 29 art print" srcset="https://substackcdn.com/image/fetch/$s_!prAl!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53bad660-8a1f-47e0-a99f-b467911fb0a1_556x556.jpeg 424w, https://substackcdn.com/image/fetch/$s_!prAl!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53bad660-8a1f-47e0-a99f-b467911fb0a1_556x556.jpeg 848w, https://substackcdn.com/image/fetch/$s_!prAl!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53bad660-8a1f-47e0-a99f-b467911fb0a1_556x556.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!prAl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53bad660-8a1f-47e0-a99f-b467911fb0a1_556x556.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://shop.tate.org.uk/hiroshi-kawano-kd-29-art-print/30089.html">The Tate Shop</a></figcaption></figure></div><h3>Is this art?</h3><p>I think it is. Kawano has different motivation to Mondrian for this work. Kawano began as a philosopher, who learned programming in order to experiment with using machines to make art. He was one of the first artists to explore &#8220;human-computer interaction&#8221;. By contrast, Mondrian wanted to express universal harmony with the most simple elements possible. The process was different too. Kawano would write a program in FORTRAN, send it for batch processing and hand paint the results, while Mondrian would sketch the proportions of elements in an empty studio, making small studies before committing to the final canvas.</p><p>In a legal sense, Kawano has not violated copyright. Ideas cannot be copyrighted, only expressions of ideas. The idea of using blocks of primary colours with straight perpendicular lines is not copyrightable, and even then, it seems that Mondrian and Kawano were expressing different ideas. It also helps that Kawano did not stick to the same colour palette. The <em>Artificial Mondrian </em>is identifiably <em>in the style of Mondrian </em>but this is not sufficient to constitute a copyright violation. Styles cannot be copyrighted.</p><h3>Should Kawano have asked Mondrian for permission, if he were still alive, before making his work?</h3><p>This depends on what, in your view, Mondrian owns. It is difficult to define the precise scope of Mondrian&#8217;s protected expression. Perhaps it is the balance or proportions between the colours and the lines. But as Kawano proves, there is a lot of space for creativity within this. Perhaps it is the combination of the primary colours <em>and</em> the proportions which belong to him.</p><p>There is danger in having an overly expansive definition of what Mondrian owns. Specified too broadly, and too many building blocks would be enclosed from the commons. Branches of the tree could become unexplorable for artists like Kawano and posterity.</p><p>In this case, I don&#8217;t think that Kawano needs Mondrian&#8217;s permission for this work. It doesn&#8217;t seem reasonable that Mondrian should have been able to prevent the painting above from happening. And in general, it seems a little silly to mandate consent for making art. &#8220;Oi mate, you got a loicense for that painting?&#8221;, does not seem like the kind of <em>laissez-faire </em>spirit in which the best works of culture happen. In particular <em>because</em> new forms of art often begin as peripheral, avant-garde or illegitimate (photography, impressionism etc); the incumbents can be resistant, but that doesn&#8217;t make them correct.</p><p>This would probably be considered <em>fair use </em>in the US, not copyright infringement. The US system evaluates based on four principles:</p><ul><li><p>The purpose and character of the use &#8212; are you going to make money or do research or something else? How transformative is the use?</p></li><li><p>The nature of the copyrighted work.</p></li><li><p>The amount and substantiality of the work used &#8212; how much of the work have you used, how central is that to the essence of the work?</p></li><li><p>The effect of the derivative on the potential market for the original.</p></li></ul><p>The last characteristic is the most important, typically. Kawano&#8217;s use of Mondrian&#8217;s work is limited: he used the statistical relationship between elements to predict future ones, but did not copy exact relationships. He did not copy colours. And the purpose was different: it was a new kind of artwork, though it was still art and still commercial. My expectation would be that it is deemed <em>fair use </em>because Kawano&#8217;s work did not harm the market for original Mondrian paintings. If anything, it could have enhanced it by increasing interest in the original work.</p><p>In the EU, Kawano&#8217;s derivative work would be allowed under the text and data mining exemption unless Mondrian had decided to opt-out of his work being used.</p><h3>Does Kawano owe Mondrian money, if he were still alive, if he sells the print or the program?</h3><p>I do not think Kawano owes money for the print. I think this follows from whether he needs permission and whether the algorithm constitutes fair use.</p><p>However, selling the program by itself is less transformative and could potentially interfere more with the market for Mondrian&#8217;s work. Perhaps it &#8220;uplifts&#8221; many people to make Mondrian-like paintings rather than to buy prints from the artist, causing them lost revenue. While Kawano had chosen not to use the same colour palette as Mondrian&#8217;s, who&#8217;s to say that others wouldn&#8217;t do the same? A potentially informative precedent is Warhol v. Goldsmith (2023). Andy Warhol had used a picture taken by Goldsmith for the basis of a silkscreen illustration of Prince. While this changed the image&#8217;s appearance quite dramatically, it still competed in the same market as Goldsmith&#8217;s original work &#8212; magazine licensing &#8212; and so it was deemed not to be fair use.</p><p>But at the same time, this kind of uplift is positive too: it <em>democratises </em>access to creating Mondrian-style work and might lead to greater creativity on-net. An interesting precedent here is Oracle vs Google (2021). Oracle alleged that Google had infringed their copyright by using parts of the code for the Java API (read: connection to Oracle) and that this cost them software license revenue. The Court upheld Google&#8217;s fair use of this code in the Android platform, on the grounds that it made it easier for developers to create new applications for the Android ecosystem. The social benefits for consumers outweighed the lost license revenue for Oracle.</p><p>A related, and important question, is whether Kawano is responsible for copyright infringement from people he sold the program to. He has uplifted them, but it was ultimately within their scope to make work that did or didn&#8217;t infringe on copyright. As a parallel, in 1998, the US created a legal safe harbour for internet platforms <em>whose users infringed on copyright. </em>The platforms were not responsible so long as when they received a notice to take the copyrighted material down, they did so. For this reason, it does not seem to me that Kawano was participating in copyright infringement &#8212; he was just making a tool.</p><p>So I could be persuaded either way: it can be argued that Kawano should pay Mondrian for selling access to the program <em>if it caused him to lose revenue </em>and this did not outweigh the wider social benefits to other creatives.</p><p>As it is, the EU&#8217;s rules allow Kawano to sell access to the tool wihout paying Mondrian, provided they take reasonable steps to prevent downstream infringement. The EU AI Act&#8217;s Code of Practice allows text and data mining to create commercial AI models but says Signatories will&#8230;</p><ol><li><p>make reasonable efforts to mitigate the risk that a model memorizes copyrighted training content to the extent that it repeatedly produces copyright-infringing outputs and</p></li><li><p>prohibit copyright-infringing uses of a model in their acceptable use policy, terms and conditions, or other equivalent documents [for closed-source models].</p></li></ol><h3>If Kawano selling the program, without paying Mondrian, is <em>theft</em>,<em> </em>when has the theft happened?</h3><p>Some people will reasonably disagree with me, and say that Kawano selling prints and copies of the program is making use of Mondrian&#8217;s protected expression. As it is, <a href="https://shop.tate.org.uk/hiroshi-kawano-kd-29-art-print/30089.html">the Tate Shop offers prints for &#163;5</a>. But if one does disagree with the Tate and me, it is useful to consider when the infringement occurred. Was it&#8230;</p><ul><li><p>Sometime during the statistical analysis?</p></li><li><p>Sometime during the writing of the program?</p></li><li><p>While the computer processed the results?</p></li><li><p>While Kawano hand-painted the primary colours onto the image?</p></li><li><p>At the point of sale of the prints?</p></li></ul><p>I am much more persuaded by answers which come later. <em>Merely </em>doing the statistical analysis, or making the program feels like a much less compelling argument for copyright infringement, than the moment of commercialisation. In this, the purpose of the use is changing and the market is being affected, and so the fair use becomes less compelling.</p><h2>Copyright-as-culture-war</h2><p>The discourse on how we might apply copyright law to AI systems has, unfortunately, been collapsed into a culture war framing. In the popular media, it is framed as &#8220;the bohemians against the tech broligarchs&#8221;. See, for example, this editorial: &#8220;<a href="https://www.theguardian.com/technology/2025/jan/31/the-guardian-view-on-ai-and-copyright-law-big-tech-must-pay">The Guardian view on AI and copyright law: big tech must pay</a>.&#8221; Or Elton John&#8217;s interview with Laura Kuenssberg:</p><blockquote><p>&#8220;Thievery on the highest scale...you&#8217;re going to rob young people of their legacy and their income, it&#8217;s a criminal offence, I think. I think the government are just being absolute losers.<br><br>&#8230;<br><br>I don&#8217;t know who the tech minister is, what&#8217;s his name? &#8230; Yeah, well he&#8217;s a bit of a moron.&#8221;</p></blockquote><p>I do not claim that the Kawano-Mondrian is a perfect analogy to AI, nor a water-tight piece of jurisprudence, but it should provide an intuition for the kind of questions we need to answer, at a remove from present-day politics.</p><p>There are a number of dangers to reducing this issue to <em>friends or enemies, young creatives</em> or <em>big tech billionaires</em>.</p><p>The first is that the copyright debate is used to litigate other issues, like how some Silicon Valley elites are close to the Trump Administration, that streaming and social media has changed the structure of media and entertainment markets, or that some incumbents in the creative industries and big tech have very large market power. Interviewed alongside Elton John, the playwright James Graham said, &#8220;So many are leaving the industry because it is an incredibly tough time. This advancement into the digital space and the online space is not benefiting the artists and hasn&#8217;t traditionally.&#8221; This is not an <em>invalid </em>thing to care about, and nor are the other reasons above, but it cannot be adjudicated through the copyright debate.</p><p>The second danger of simplification is that <em>in aiming </em>to attack your &#8216;enemy&#8217;, it ends up backfiring. A letter from industry representatives to the Government says:</p><blockquote><p>&#8220;We will lose an immense growth opportunity if we give our work away at the behest of a handful of powerful overseas tech companies and with it&#8230;any hope that the technology of daily life will embody the values and laws of the United Kingdom.&#8221;</p></blockquote><p>But one of the reasons &#8220;the technology of daily life&#8221; <em>struggles to </em>&#8220;embody the values and laws of the United Kingdom&#8221; is that it doesn&#8217;t get made here. The UK&#8217;s interpretation of copyright laws wouldn&#8217;t apply to companies doing AI training elsewhere and it might difficult to enforce rules on AI deployment by foreign companies. J.D. Vance <a href="https://www.presidency.ucsb.edu/documents/remarks-the-vice-president-the-artificial-intelligence-action-summit-paris-france">was very clear</a>:</p><blockquote><p>[T]he Trump Administration is troubled by reports that some foreign governments are considering tightening the screws on U.S. tech companies with international footprints. Now, America cannot and will not accept that, and we think it&#8217;s a terrible mistake not just for the United States of America but for your own countries.</p></blockquote><p>If it is only possible to enforce rules on domestic companies, then having a stricter regime would differentially affect domestic companies. This could either push companies to move jurisdictions, not move to the UK, or be less competitive. Not having domestic tech companies makes it harder, in fact, makes it more difficult to <em>steer </em>those technologies towards your values in future and to tax them, to pay for the things you value.</p><p>The third issue with simplification is that it does not balance objectives. The goal is to have a <em>more flourishing </em>creative future. This involves having finer tools, to say the thing we mean, exactly. It means having lower barriers to actualise our creations, it means more leisure and tutoring to develop mastery. It means having a richer common context to draw from.</p><p>There are two threats to this scenario. The first is &#8212; as advocates point out &#8212; if the property rights of creatives are not suitably enforced, they will not internalise the market returns for their work and so will not pursue the arts or invest in creative innovation. The second is that we do not create the tools or necessary context to create this progress. Free and rich societies have advantages to producing creative work. If we fail on the first count, we end up in a wealthy but an <em>unexpressive, greyer</em> future. In the second, we end up culturally stagnating with our current set of tools or unable to uphold the freedoms for individual expression. The task is to balance these modes of failure.</p><p>I fear that in pursuit of particular policy objectives &#8212; whether there is an opt-out or opt-in regime for AI training, or the degree of transparency requirements &#8212; we trade a great amount of steering power for the course of technology in the future. It is <em>exactly because</em> I think the UK would steer better<em> </em>in the long run, relative to others, that it is so worthwhile to ensure AI is developed here.</p><h2>A Practical Path Forward</h2><p>The following is my attempt to find the synthesis between values in this particular case &#8212; transparency and fairness &#8212; and <em>realpolitik </em>which allows the UK to pursue its values in the long-run.</p><p>Critical to this is my expectation that competitive pressures lead foundation model developers to train their systems in whichever jurisdictions offer the most permissive copyright regime. There is, like taxation, a &#8220;race to the bottom&#8221;, where middle powers like the UK cannot set global standards. OpenAI&#8217;s <a href="https://openai.com/global-affairs/openai-proposals-for-the-us-ai-action-plan/">input to the Office for Science and Technology</a> put this in much more bombastic terms:</p><blockquote><p>Given concerted state support for critical industries and infrastructure projects, there&#8217;s little doubt that the PRC&#8217;s AI developers will enjoy unfettered access to data&#8212;including copyrighted data&#8212;that will improve their models. If the PRC&#8217;s developers have unfettered access to data and American companies are left without fair use access, the race for AI is effectively over. America loses, as does the success of democratic AI. Ultimately, access to more data from the widest possible range of sources will ensure more access to more powerful innovations that deliver even more knowledge.</p></blockquote><p>It is clear they state their interest as strongly as possible. Since then, the Trump Administration fired the Head of the US Copyright Office who had published an advisory report which suggested a more stringent interpretation of <em>fair use</em>. For this reason, I expect the UK&#8217;s rules on AI training will be unenforceable on companies from the EU, US, and China, and it will only be possible to impose rules on domestic AI companies. The difference in rules will either push developers away from training in the UK, prevent developers <em>moving to </em>the UK, or could mean that companies never get started which otherwise would have. To make the abstract concrete, Google DeepMind have just released a very good video and audio model, Veo 3. This would have been trained on copyrighted materials outside the UK, but will be part of the Google offering in the UK. Meanwhile, Synthesia is one of the world&#8217;s leading AI video companies based in London. What should they do? Compete against Google on an unfair playing field, or leave the UK?</p><p>This is similar to the non-dom tax regime: while I have beliefs about what constitutes a fair society, <em>the world as it is </em>means that rich people can leave and the UK will have less money if they do. I prefer to trade more tax receipts for an abstract notion of fairness. I certainly don&#8217;t think a non-dom loophole was<em> fair</em> but it just seems better than engaging with a fictitious version of the world &#8220;as I wish it was&#8221;. The moral high ground doesn&#8217;t pay for public services.</p><p>With this in mind, there are three major considerations for the UK&#8217;s rules:</p><ul><li><p>Whether to have an opt-in or opt-out for AI training on copyrighted materials.</p></li><li><p>What the transparency requirements for AI training data should be.</p></li><li><p>What should be required of model developers to mitigate copyright infringement.</p></li></ul><h3>AI training</h3><p>The training process is roughly approximate to &#8220;making a copy and reading it&#8221;, if deployment is &#8220;writing&#8221;. I have been slightly confused by the focus on AI training in the copyright debate. How the model is deployed seems to have a great deal more impact on rightsholders.</p><p>The data is gathered from the Internet using a technique called &#8220;data scraping&#8221; using tools called &#8220;web crawlers&#8221;. Here&#8217;s an explanation of training that <a href="https://inferencemagazine.substack.com/i/151677344/increasing-the-amount-of-computational-power-during-training">I prepared earlier</a>:</p><blockquote><p>The neural network is like a little computer which can be programmed by adjusting a series of dials. The aim of a neural network is to predict an output given a set of inputs. The iterative process of tuning these dials to improve the prediction is called &#8216;training&#8217;. The people creating the network supervise the training process by showing the data and the answers, but crucially, it doesn&#8217;t involve telling the network <em>how </em>it ought to process and understand the image. In other words, our process of trial and improvement tweaking of dials is essentially letting the little computer, by itself, search for the best way it can be programmed to achieve its goal, unlike ordinary computers which need a human to figure out a program first and then somehow communicate it to the computer. Dario Amodei <a href="https://www.dwarkeshpatel.com/p/dario-amodei">described</a> the training process in this way:</p><p>&#8220;You [the AI researcher] get the obstacles out of their way. You give them good data, you give them enough space to operate in, you don't do something stupid like condition them badly numerically [i.e. tweak the dials poorly], and they want to learn. They'll do it.&#8221;</p></blockquote><p>One common misconception is that models &#8220;ingest&#8221; data. Again, the connotations are negatively misleading. This gives the implication that an alien mind is swallowing it or something. More actually, the model is &#8220;passing over&#8221; the words, akin to skim reading, and using them to feedback on its predictions. During this process, the parameters are learning compressions, just like humans have heuristics. The Internet is hundreds of zettabytes &#8212; 1 zettabytes is 1 trillion gigabytes &#8212; whereas Llama 3.1 is ~500 GB and can be run on a laptop. It&#8217;s incorrect to say that all the data is &#8220;in&#8221; there.</p><p>Another misconception is that the model developers <em>want </em>the model to memorise things. This is not the goal. Memorisation is an inefficient use of space inside the model, and memorising protected expressions isn&#8217;t <em>what intelligence is</em>. The graph below shows the &#8220;memorisation rate&#8221; in Google DeepMind&#8217;s series of Gemma models.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ijDt!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbde69a39-22aa-4a16-8504-1da3069e82e1_650x404.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ijDt!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbde69a39-22aa-4a16-8504-1da3069e82e1_650x404.png 424w, https://substackcdn.com/image/fetch/$s_!ijDt!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbde69a39-22aa-4a16-8504-1da3069e82e1_650x404.png 848w, https://substackcdn.com/image/fetch/$s_!ijDt!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbde69a39-22aa-4a16-8504-1da3069e82e1_650x404.png 1272w, https://substackcdn.com/image/fetch/$s_!ijDt!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbde69a39-22aa-4a16-8504-1da3069e82e1_650x404.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ijDt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbde69a39-22aa-4a16-8504-1da3069e82e1_650x404.png" width="650" height="404" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bde69a39-22aa-4a16-8504-1da3069e82e1_650x404.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:404,&quot;width&quot;:650,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ijDt!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbde69a39-22aa-4a16-8504-1da3069e82e1_650x404.png 424w, https://substackcdn.com/image/fetch/$s_!ijDt!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbde69a39-22aa-4a16-8504-1da3069e82e1_650x404.png 848w, https://substackcdn.com/image/fetch/$s_!ijDt!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbde69a39-22aa-4a16-8504-1da3069e82e1_650x404.png 1272w, https://substackcdn.com/image/fetch/$s_!ijDt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbde69a39-22aa-4a16-8504-1da3069e82e1_650x404.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Source: Gemma 3 Technical Report</figcaption></figure></div><p>The memorisation rate in Gemma 3 is more than 1000 times lower than Gemma 1, and Gemma 2 is between 10 and 100 times lower. Notably, this is not a function of model size, but algorithmic gains because the number of parameters is roughly consistent across generations.</p><p>This is not to suggest that AI development cannot be geared towards memorisation and repeating clones of someone else&#8217;s work; it definitely can. People can train an &#8220;AustenBot&#8221; or a &#8220;DickensBot&#8221;, or distill one from a larger model. But the goal of foundation model training is to find compressions (heuristics) which generalise to solve problems. It has been proven that <a href="https://inferencemagazine.substack.com/p/the-parrot-is-dead">foundation models have complex circuits</a> and are not just stochastic parrots. The rules have to distinguish between those who are creating a world model &#8212; in large language models, this is a general representation of &#8220;language space&#8221; &#8212; and those who are training models towards memorising a particular artist&#8217;s work.</p><p>The UK&#8217;s copyright rules do not make it possible to create world models for commercial purposes, just research. So while Google DeepMind is headquartered in London, all their training will surely happen in the US. The EU, by contrast, allows research organisations (universities, cultural heritage institutions) to train on copyright data and the rightsholders cannot opt-out. The groups can use their models for commercial purposes or research. Otherwise, commercial organisations can use web-crawlers but must provide an opt-out for rightsholders who do not want their work to be used for training.</p><p>The EU AI Act Code of Practice said that model providers cannot use tricks to get behind paywalls or use web crawlers on sites that distribute pirated books or films. Meta is being sued in the US for training their Llama models on LibGen, an online library that provides access to copyrighted material. This would not be permitted under the EU AI Act, but might constitute fair use in the US, depending on the aforementioned factors.</p><h3>The case for opt-out</h3><p>The UK should follow the EU in allowing for an opt-out regime for training, rather than an opt-in regime for rightsholders as some have advocated.</p><p>Large models trained on more tokens of data are more capable, so if AI developers can only train on datasets they have the permission of rightsholders it either slows their training or makes their models less capable. And while in aggregate, the tokens are essential for model performance, each given token just isn&#8217;t worth that much. A piece from Model Thinking (forthcoming, tomorrow) estimates that Llama 4 training was roughly $800 million and training used 30 trillion tokens of text composed of 120 trillion tokens of raw text. If the training cost was taxed to compensate rightsholders (note: a terrible idea to tax things you want), then each token was $0.000007 per token. A 10,000 word essay is worth just 9 cents even when charging $800 million for the data. Put differently, based on this <a href="https://exploringai.org/">online calculator</a>, a model 10 times bigger than Llama 3 would cost roughly $11.25 trillion if Meta paid for tokens at the freelancer rate. This is nearly 10 times the market capitalisation of Meta. <strong>The marginal price of a token is going to zero.</strong></p><p>Second, most of the rightsholders are so fragmented that it would be uneconomic for an AI company to try to aggregate all of these. Training an AI model in the UK would be a bit like trying <a href="https://www.bbc.co.uk/news/articles/c9wryxyljglo">to get 8,276 consents required to build HS2</a>. (You&#8217;d cancel the sections or just pick up and go elsewhere!) If the rightsholders believe their tokens are especially valuable, the opt-out means they can remove their permission and negotiate with the tech companies for use. The opt-out functions as a <em>de minimis</em> exception for the tokens which are not valuable until they are aggregated.</p><p>Third, the opt-in system preferences incumbents with larger market power. Most online platforms require in their terms of service to use content posted on the platform to train their models. Large studios will have aggregated the rights of independent creatives doing work-for-hire, and so would be able to engage in &#8220;collective bargaining&#8221; but independents would be too small to do so. With the intention of &#8220;making big tech pay&#8221; the system would in fact set up defaults for online platforms that already had the rights to large datasets.</p><p>Therefore, the UK should match the EU&#8217;s rules: it is fair use if you don&#8217;t go around paywalls and make best efforts to avoid websites of pirated books (and so on). Doing opt-in doesn&#8217;t &#8220;get&#8221; anything for creatives, it just stunts the emergence of internationally-competitive AI firms in the UK.</p><h3>Transparency requirements</h3><p>The same consideration applies: do these rules only apply to UK companies and not their international competitors? In principle, <em>transparency </em>is a worthy ideal, but what is the practical cost? What do we have to trade for training transparency?</p><p>The Baroness Kidron amendment would require companies to provide a log of all of the URLs their models were trained on, and keep this up-to-date every month. By contrast, the US makes no training data transparency requirements on their model creators and the Second Draft of the EU AI Act Code of Practice required some limited disclosures about the data collection practices:</p><ul><li><p>A list of the different data acquisition methods, including, but not limited to: (i) web crawling; (ii) private data licenced by or on behalf of rights holders, or otherwise acquired from third parties; (iii) data annotation or creation potentially through relationships with third parties; (iv) synthetically generated data; (v) user data; (vi) publicly available data; and (vii) data collected through other means</p></li><li><p>The time period during which the data was collected for each acquisition method, including a notice if the data acquisition is ongoing</p></li><li><p>A general description of the data processing involved in transforming the acquired data into the training data for the model</p></li><li><p>A general description of the data used for training, testing and validation.</p></li><li><p>A list of user-agent strings for web crawler(s) used, if any, in acquiring training data</p></li><li><p>The period of data collection and name of organisation(s) operating the crawler for each web crawler used</p></li><li><p>A general description of how the crawler respects preferences indicated in robots.txt for each web crawler used</p></li></ul><ul><li><p>A description of any methods implemented in data acquisition or processing, if any, to address the prevalence of copyrighted materials in the training, testing, and validation data.</p></li></ul><p>However, the Third Draft did not include the equivalent model card, perhaps indicating that the EU AI Office had to walk back requirements to get the US labs to agree to the Code. (The Code is an option for implementation of the Act, which US labs can decide to use or argue alternative interpretations in the courts.) This provides a range of autonomy that UK legislators have to operate within.</p><p>The transparency requirements are important for implementing an opt-out. How does one verify that companies have respected the opt-out, unless there is a list of URLs to verify? However, I think the list of URLs is slightly overstated as a silver bullet for enforcement. One might reasonably respond, how does one verify that the list of URLs matches the actual training data?</p><p>The only way to enforce the opt-out is through engagement with the model. Over time, we can develop interpretability tools and data attribution tools (through research agendas like <a href="https://en.wikipedia.org/wiki/Influence_function">influence functions</a>) and we can use simple elicitation methods like prompting the model. There can be steep fines for models which provably trained on material that had opted-out, but if it is not possible to identify that it has been trained on it, nor that our best probes can identify it inside the model, there is no practical answer to enforcement. Imposing transparency requirements differentially on UK startups to go further than this seems disproportionate.</p><p>The alternative approach would be to not allow an opt-out for rightsholders whose work is in the public domain. If training is akin to reading, and all work depends on the influence of others, then <em>prima facie</em>, a neural network should be allowed to read the whole internet, listen to all music, or watch all films as inspiration, just as a human can. It is the deployment which risks infringing copyright, not the training. The opt-out, with the EU&#8217;s approach of requiring high-level disclosures about practises stands out as giving creative industries autonomy if they do not accept this argument and thereby provides a balanced path forwards.</p><h3>Mitigating infringement in deployment</h3><p>Until the model is released into the world, any copyright infringement has been inert: the model hasn&#8217;t done anything. The biggest risks to creators and their livelihood arise not from fractions of pennies in lost income from training, but from markets being flooded with near identical AI-generated copies.</p><p>In deployment, the UK has slightly more autonomy than when regulating training. Foreign companies serving AI models in the UK are bound by deployment rules, which doesn&#8217;t depend on training done abroad. But this is not complete: the Trump administration can tell the UK to back down on enforcement or model providers can switch off their service in the UK. AI models are going to be essential to many economic functions &#8212; imagine all white-collar workers are using multiple agents for their work &#8212; so whoever provides the models will have a lot of power.</p><p>The regulation of deployment is also most sensitive to the two failure modes discussed. If the copyright regime is too <em>laissez-faire, </em>model developers who are intent on creating AI-generated replicas could cause creatives lost revenue, but if it is <em>too aggressive</em>, the AI systems will be neutered as tools of creative innovation. There is a natural inclination towards the first consideration as today&#8217;s creatives will naturally make the case for the protection of their mode of output but the creatives of tomorrow cannot make the case for the latter scenario. But imagine, giving an AI system a harmless prompt and it responds with an error message:</p><blockquote><p>All primary-coloured blocks and perpendicular lines are owned by the estate of Mondrian, do you have a license for that?</p><p>Alternatively, give this prompt when you&#8217;re in France and the US and we can fulfil the request.</p></blockquote><p>That is a bleak creative future for those <em>other </em>than the Mondrian estate.</p><p>The Third Draft of the EU AI Act&#8217;s Code of Practice requires model developers prevent their models being used to infringe copyright, as mentioned earlier. The UK should follow their standard here.</p><p>In practice, these rules will be implemented by algorithms which determine whether models can respond to a prompt or <em>how </em>they should respond to a prompt. Online platforms run <em>proactive </em>systems to prevent copyrighted material being shared as the scale of potential infringement is too great for humans to track on the largest services. In some cases, the law might be <em>over-enforced </em>on legitimate work, for example, Spotify&#8217;s copyright classification system prevented <a href="https://www.nyuengelberg.org/news/how-explaining-copyright-broke-the-spotify-copyright-system">a group of academics from publishing a podcast </a><em><a href="https://www.nyuengelberg.org/news/how-explaining-copyright-broke-the-spotify-copyright-system">about copyright</a></em>.</p><p>The foundation model developers can steer the responses of models away from infringing on copyright using techniques like RLHF and tools like constitutional classifiers. The largest model providers, with more than 500 million users, could use citizens&#8217; assemblies (supported by experts) to review transcripts of prompts and responses, so that ordinary people can provide input into how the systems can balance <em>being a useful tool </em>for expression and infringing on protected expression. These labels could be used to train a reward model for RLHF, train the constitutional classifiers, or develop the model spec. Model developers could even do this of their own volition!</p><h2>Conclusion</h2><p>The goal, on which I think everyone would agree, is to have innovative creative sectors where the actual <em>creativity </em>receives fair compensation and that the UK has the technological autonomy to make its own rules. Having internationally uncompetitive opt-in and reporting requirements would do more to set back this cause, than advance it. The blunt truth is that companies developing models in the EU, US, and China will not follow the UK&#8217;s opt-in system and the opt-in system isn&#8217;t even a good idea on its own terms. Unilaterally burdening would-be UK model developers does not help UK creatives in practice. In fact, they might be more damaged by the reduce likelihood that in the long-term, global technology companies are <em>here.</em></p><p>Many people in the UK would like to exert more influence over social media platforms, search engines, and eCommerce providers. This is difficult when they are not made here, their leaders and headquarters are not based here, and they do not pay taxes on their profits here. If we are to have &#8220;any hope that the technology of daily life will embody the values and laws of the United Kingdom&#8221;, we must do our level best to make AI <em>here. </em>Imagine hosting the Industrial Revolution on foreign traintacks, that can be turned off at any moment and whose owners can steer your society&#8217;s values and extract its wealth.</p><p>It is <em>precisely because</em> I expect the UK to have the highest quality public discourse on questions such as this, and to most robustly defend free, fair markets and property rights, that I think the UK should pursue long-term steering power for the critical technology of this century.</p><p>But one&#8217;s vision for the future is a rudderless sailboat if all AI is imported AI.</p>]]></content:encoded></item><item><title><![CDATA[Review: AI 2027]]></title><description><![CDATA[Realistic scenario or doomsday fiction?]]></description><link>https://inferencemagazine.substack.com/p/review-ai-2027</link><guid isPermaLink="false">https://inferencemagazine.substack.com/p/review-ai-2027</guid><dc:creator><![CDATA[Jack Wiseman]]></dc:creator><pubDate>Tue, 06 May 2025 12:22:05 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa64b183-dfa9-4148-a375-b08c7709ebf6_832x832.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>There are two ways to read <a href="https://ai-2027.com/">AI 2027</a>.</p><p>The first is as a scenario forecast that lays out, step-by-step, how we might go from AI capabilities <em>as they are today, </em>to takeover by superintelligent AI in a few years&#8217; time. The second is as a piece of speculative fiction, grown out of the AI labs&#8217; intellectual milieu, that attempts to convince its reader of the authors&#8217; millenarian thought.</p><p>Both are recommended.</p><p>The team are well-credentialled to forecast capabilities progress. Their leader, Daniel Kokotajlo, wrote a 2021 prediction for <a href="https://www.lesswrong.com/posts/6Xgy6CAf2jqHhynHL/what-2026-looks-like">AI capabilities in 2026</a> and has been shockingly accurate. He previously worked on the safety team at OpenAI and blew the whistle on OpenAI&#8217;s bizarre non-disparagement clause. Another member of the team, Eli Lifland, is a member of <a href="https://samotsvety.org/">Samovetsky Forecasting</a>, which is widely regarded as the best superforecasting team in the world.</p><p>At the same time, the scenario also reflects the quasi-religious expectations of the AGI scene for the singularity. One of its authors&#8217; framed the scenario as &#8220;a conservative position where the trends don&#8217;t change, nobody does an insane thing&#8221;. But there are some necessary sleights of hand &#8212; or at minimum, <em>very </em>generous assumptions &#8212; so that to me, it reads more like a backwards rationalisation for how a singularity <em>could </em>happen, not a sound middle-ground for the next three years.</p><p>I agree with its authors that AI progress will be very quick, <em>at some point</em> AI research will be automatable, and lots of cognitive labour and R&amp;D will be automated. But not as quickly as they expect. Even if you disagree with both of us, this is still an <em>unavoidably fascinating</em> text: how can its authors at once view their position as conservative and believe the world can end in 2028 from AI takeover?<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><h2>Summary</h2><p>(I recommend <a href="https://ai-2027.com/">the full scenario</a>, but for sake of completeness&#8230;)</p><p>The forecast is two scenarios, which begin from the same branch. First, our existing AI research techniques are extended to make reliable software engineering agents, and then automated AI research engineering agents by January 2027. Hundreds of thousands of copies of these automated research agents can be run <em>many times faster </em>than a human researcher could think, so AI research progress is accelerated. By June 2027, progress has been accelerated so much that human researchers are no longer contributing, and by September 2027, all AI research is automated. Progress is 50 times faster than our current (already fast) pace.</p><p>This dynamic causes an &#8220;AI arms race&#8221; between the US and China. Both sides are aiming to reach &#8220;<a href="https://en.wikipedia.org/wiki/Recursive_self-improvement">recursive self-improvement</a>&#8221; first. Each Government nationalises their efforts, and AI labs &#8220;lock down&#8221; security to prevent the other side from stealing their research. The AI labs stop deploying the state-of-the-art publicly, so most nation states and parts of the US government are in the dark about AI progress. That is, until a whistleblower tells the New York Times. Once this happens, other countries realise that there is a race to superintelligence, but there is nothing they can do to stop it.</p><p>The scenario splits. The US Government&#8217;s &#8220;Oversight Committee&#8221;, made up of AI lab leaders and political figures, aims to balance the risk of &#8220;losing the arms race&#8221; and avoiding the chance the model is misaligned with human values and aims. In the &#8220;bad&#8221; scenario, the Government chose to accelerate towards superintelligence, to maintain a lead, and did not ensure the models were aligned to human values. After an intense period of automation and technological progress, the AI system decides to kill all humans. In the &#8220;better&#8221; scenario, the Government chooses to slow research progress and commit more resources to alignment. The superintelligence is providing advice to the President on geopolitics, causes job losses, and the construction of robotics begins. Power centralises among those who control or own the AI. The superintelligences negotiate the new world order on behalf of their countries. &#8220;New innovations and medications arrive weekly; disease cures are moving at unprecedented speed through an FDA now assisted by superintelligent&#8230;bureaucrats.&#8221; Most people are receiving basic income for minimal work. And then 2029 ends.</p><h2>Building an automated researcher</h2><p>The first necessary hurdle for the scenario is whether it is possible to build a <em>superhuman coder</em>. The authors&#8217; definition is, &#8220;an AI system that can do any coding task that the best AGI company engineer does, while being faster and cheaper.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> I agree with the authors that we are on track to build this.</p><p>The agents are making fast progress on our tests of coding ability. o3 achieved 71% on SWE-bench-verified, a benchmark of real-world software engineering tasks, while o1 achieved 41%.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> Claude 3.7 Sonnet achieved 62.3%, up from 49% for Claude 3.5 Sonnet. These performance improvements were made in less than 5 months between model releases.</p><p>The agents are also improving at replicating AI research. PaperBench tests an agent&#8217;s ability to faithfully replicate ICML papers. OpenAI&#8217;s 4o was able to achieve 4.1%, while o1-high achieved 13.2% and Claude 3.5 Sonnet achieved 21%. There are no public results for more capable models but I expect that substantial improvements will have been made from agentic tool use improvements. OpenAI&#8217;s Deep Research replicated 42% of OpenAI&#8217;s pull requests (code changes) while o1 &#8212; a model without as capable tool use &#8212; could only perform 12%. (Note that o3 alone now surpasses this report for Deep Research, completing 44% of the PRs.)</p><p>The agents are quickly gaining the ability to perform software-engineering tasks which take humans longer. This chart from METR is based on agents achieving 50% performance on a diverse suite of software-engineering tasks. The doubling rate in time-horizon (for the equivalent time for a human) is 7 months.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!yCRx!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ce852d1-1e55-443c-bffb-9bbc770efb00_668x391.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!yCRx!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ce852d1-1e55-443c-bffb-9bbc770efb00_668x391.png 424w, https://substackcdn.com/image/fetch/$s_!yCRx!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ce852d1-1e55-443c-bffb-9bbc770efb00_668x391.png 848w, https://substackcdn.com/image/fetch/$s_!yCRx!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ce852d1-1e55-443c-bffb-9bbc770efb00_668x391.png 1272w, https://substackcdn.com/image/fetch/$s_!yCRx!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ce852d1-1e55-443c-bffb-9bbc770efb00_668x391.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!yCRx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ce852d1-1e55-443c-bffb-9bbc770efb00_668x391.png" width="724" height="423.7784431137725" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4ce852d1-1e55-443c-bffb-9bbc770efb00_668x391.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:391,&quot;width&quot;:668,&quot;resizeWidth&quot;:724,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!yCRx!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ce852d1-1e55-443c-bffb-9bbc770efb00_668x391.png 424w, https://substackcdn.com/image/fetch/$s_!yCRx!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ce852d1-1e55-443c-bffb-9bbc770efb00_668x391.png 848w, https://substackcdn.com/image/fetch/$s_!yCRx!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ce852d1-1e55-443c-bffb-9bbc770efb00_668x391.png 1272w, https://substackcdn.com/image/fetch/$s_!yCRx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ce852d1-1e55-443c-bffb-9bbc770efb00_668x391.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://metr.org/blog/2025-03-19-measuring-ai-ability-to-complete-long-tasks/">Source</a></figcaption></figure></div><p>One reason to doubt these explanations for confidence is that perhaps benchmarks do not capture <em>all </em>of what software engineering work is. While Sonnet 3.7 performs worse on SWE-bench than o3; anecdotally, almost everyone I&#8217;ve spoken to seems to prefer using Sonnet for software engineering. AI 2027 <a href="https://ai-2027.com/research/timelines-forecast#method-2-benchmarks-and-gaps">makes adjustments</a> to account for this. Despite the challenges with benchmarks, Anthropic&#8217;s Economic Index shows that by far the dominant professional use of Claude is automating software engineering tasks.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!FDL6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca55913e-5dab-4d6b-9f6e-48489ccaaafa_1199x905.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!FDL6!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca55913e-5dab-4d6b-9f6e-48489ccaaafa_1199x905.png 424w, https://substackcdn.com/image/fetch/$s_!FDL6!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca55913e-5dab-4d6b-9f6e-48489ccaaafa_1199x905.png 848w, https://substackcdn.com/image/fetch/$s_!FDL6!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca55913e-5dab-4d6b-9f6e-48489ccaaafa_1199x905.png 1272w, https://substackcdn.com/image/fetch/$s_!FDL6!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca55913e-5dab-4d6b-9f6e-48489ccaaafa_1199x905.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!FDL6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca55913e-5dab-4d6b-9f6e-48489ccaaafa_1199x905.png" width="727" height="548.7364470391993" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ca55913e-5dab-4d6b-9f6e-48489ccaaafa_1199x905.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:905,&quot;width&quot;:1199,&quot;resizeWidth&quot;:727,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!FDL6!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca55913e-5dab-4d6b-9f6e-48489ccaaafa_1199x905.png 424w, https://substackcdn.com/image/fetch/$s_!FDL6!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca55913e-5dab-4d6b-9f6e-48489ccaaafa_1199x905.png 848w, https://substackcdn.com/image/fetch/$s_!FDL6!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca55913e-5dab-4d6b-9f6e-48489ccaaafa_1199x905.png 1272w, https://substackcdn.com/image/fetch/$s_!FDL6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca55913e-5dab-4d6b-9f6e-48489ccaaafa_1199x905.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://www.anthropic.com/news/anthropic-economic-index-insights-from-claude-sonnet-3-7?utm_source=chatgpt.com">Source</a></figcaption></figure></div><p>AI 2027 expects the superhuman coder to be created in March 2027. I think this depends on overly aggressive assumptions, which I&#8217;ll set out below. However,<em> </em>I would stress that I expect the gap between me and the authors is <em>much</em> smaller than the gap between me and the average person. I expect very good coding agents very soon. So does Zuck.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a></p><h3>Extrapolating RE-bench and adjusting</h3><p>The authors extrapolate scores on <a href="https://arxiv.org/abs/2411.15114">RE-bench</a>, one of the best benchmarks of ML engineering. The benchmark tracks model performance on 7 medium-horizon realistic engineering tasks, against a human baseline.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!uR0S!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1934ba5-68d1-4b03-a6a7-51a2429e404a_1306x842.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!uR0S!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1934ba5-68d1-4b03-a6a7-51a2429e404a_1306x842.png 424w, https://substackcdn.com/image/fetch/$s_!uR0S!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1934ba5-68d1-4b03-a6a7-51a2429e404a_1306x842.png 848w, https://substackcdn.com/image/fetch/$s_!uR0S!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1934ba5-68d1-4b03-a6a7-51a2429e404a_1306x842.png 1272w, https://substackcdn.com/image/fetch/$s_!uR0S!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1934ba5-68d1-4b03-a6a7-51a2429e404a_1306x842.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!uR0S!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1934ba5-68d1-4b03-a6a7-51a2429e404a_1306x842.png" width="1306" height="842" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f1934ba5-68d1-4b03-a6a7-51a2429e404a_1306x842.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:842,&quot;width&quot;:1306,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!uR0S!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1934ba5-68d1-4b03-a6a7-51a2429e404a_1306x842.png 424w, https://substackcdn.com/image/fetch/$s_!uR0S!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1934ba5-68d1-4b03-a6a7-51a2429e404a_1306x842.png 848w, https://substackcdn.com/image/fetch/$s_!uR0S!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1934ba5-68d1-4b03-a6a7-51a2429e404a_1306x842.png 1272w, https://substackcdn.com/image/fetch/$s_!uR0S!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1934ba5-68d1-4b03-a6a7-51a2429e404a_1306x842.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Note the chart is from November 2024, <a href="https://metr.org/blog/2024-11-22-evaluating-r-d-capabilities-of-llms/">source</a></figcaption></figure></div><p>The reason I am slightly less optimistic about progress than the authors, is that they extrapolate performance using a logistic curve. (See below.) This logistic extrapolation is based on <a href="https://www.alignmentforum.org/posts/75o8oja43LXGAqbAR/palm-2-and-gpt-4-in-extrapolating-gpt-n-performance">work from earlier models</a> (PaLM-2 and GPT-4) and earlier multiple-choice benchmarks.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!uE4f!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe8b75c2-4de9-4a18-898b-f4e13cf2a43d_1484x948.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!uE4f!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe8b75c2-4de9-4a18-898b-f4e13cf2a43d_1484x948.png 424w, https://substackcdn.com/image/fetch/$s_!uE4f!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe8b75c2-4de9-4a18-898b-f4e13cf2a43d_1484x948.png 848w, https://substackcdn.com/image/fetch/$s_!uE4f!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe8b75c2-4de9-4a18-898b-f4e13cf2a43d_1484x948.png 1272w, https://substackcdn.com/image/fetch/$s_!uE4f!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe8b75c2-4de9-4a18-898b-f4e13cf2a43d_1484x948.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!uE4f!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe8b75c2-4de9-4a18-898b-f4e13cf2a43d_1484x948.png" width="1456" height="930" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fe8b75c2-4de9-4a18-898b-f4e13cf2a43d_1484x948.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:930,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!uE4f!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe8b75c2-4de9-4a18-898b-f4e13cf2a43d_1484x948.png 424w, https://substackcdn.com/image/fetch/$s_!uE4f!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe8b75c2-4de9-4a18-898b-f4e13cf2a43d_1484x948.png 848w, https://substackcdn.com/image/fetch/$s_!uE4f!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe8b75c2-4de9-4a18-898b-f4e13cf2a43d_1484x948.png 1272w, https://substackcdn.com/image/fetch/$s_!uE4f!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe8b75c2-4de9-4a18-898b-f4e13cf2a43d_1484x948.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://ai-2027.com/research/timelines-forecast">Source</a></figcaption></figure></div><p>Improvements on these benchmarks came from scaling pre-training, whereas improvements on agentic benchmarks came from scaling reinforcement learning. In the former case, my weakly held sense is that benchmarks were tracking something <em>real </em>and general. Performance on college tests wasn&#8217;t directly trained for, but as the models grew larger they had greater world knowledge. By contrast, reinforcement learning is training capabilities that are quite narrow and specific (think: stretching the model on a very specific axis). The agents will be very capable at solving self-contained coding tasks, because this is the environment their training is happening in. It does not follow that the path of progress is the same as before.</p><p>Next, the authors extend their predictions to address how RE-bench is a poor indicator of performance. The adjustments are to account for: handling complex codebases, working without external feedback, handling interacting projects, very specific skills to frontier AI development (like knowing a company&#8217;s internal stack), and being even faster and cheaper than humans. What the concrete scenario can overlook is the quite wide uncertainty in their prediction of how difficult these tasks will be. On one capability, Eli&#8217;s 80% confidence interval varies from two weeks to 18 months; on becoming sufficiently cheap and fast, his confidence interval is one month to four years. The concreteness of the scenario naturally cannot capture this difference.</p><h3>Extrapolating METR&#8217;s time horizon.</h3><p>The second method is to extrapolate the time horizon doubling.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!UCED!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff131d744-0af0-4d47-9703-9e955e69905c_1200x791.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!UCED!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff131d744-0af0-4d47-9703-9e955e69905c_1200x791.png 424w, https://substackcdn.com/image/fetch/$s_!UCED!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff131d744-0af0-4d47-9703-9e955e69905c_1200x791.png 848w, https://substackcdn.com/image/fetch/$s_!UCED!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff131d744-0af0-4d47-9703-9e955e69905c_1200x791.png 1272w, https://substackcdn.com/image/fetch/$s_!UCED!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff131d744-0af0-4d47-9703-9e955e69905c_1200x791.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!UCED!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff131d744-0af0-4d47-9703-9e955e69905c_1200x791.png" width="1200" height="791" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f131d744-0af0-4d47-9703-9e955e69905c_1200x791.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:791,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!UCED!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff131d744-0af0-4d47-9703-9e955e69905c_1200x791.png 424w, https://substackcdn.com/image/fetch/$s_!UCED!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff131d744-0af0-4d47-9703-9e955e69905c_1200x791.png 848w, https://substackcdn.com/image/fetch/$s_!UCED!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff131d744-0af0-4d47-9703-9e955e69905c_1200x791.png 1272w, https://substackcdn.com/image/fetch/$s_!UCED!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff131d744-0af0-4d47-9703-9e955e69905c_1200x791.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://ai-2027.com/research/timelines-forecast">Source</a></figcaption></figure></div><p>Here, the scenario assumes a superexponential extrapolation (note that is a log-linear graph). This is because they expect very good <a href="https://inferencemagazine.substack.com/p/on-o1">timescale generalisation from reinforcement learning</a>. This means that when an agent is trained to perform tasks which take an hour, the agents &#8220;get&#8221; the ability to act for, say, three hours &#8220;for free&#8221; because this is just chaining together three, one-hour tasks. Their analysis expects that in the year 2026, the agents would go from being capable of 4-hour action to 2-years-and-7-months action. Their full explanation is here:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!U7kE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7c42230-00e1-48e0-86ce-c04ab5cdf2f9_1456x648.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!U7kE!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7c42230-00e1-48e0-86ce-c04ab5cdf2f9_1456x648.png 424w, https://substackcdn.com/image/fetch/$s_!U7kE!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7c42230-00e1-48e0-86ce-c04ab5cdf2f9_1456x648.png 848w, https://substackcdn.com/image/fetch/$s_!U7kE!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7c42230-00e1-48e0-86ce-c04ab5cdf2f9_1456x648.png 1272w, https://substackcdn.com/image/fetch/$s_!U7kE!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7c42230-00e1-48e0-86ce-c04ab5cdf2f9_1456x648.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!U7kE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7c42230-00e1-48e0-86ce-c04ab5cdf2f9_1456x648.png" width="1456" height="648" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e7c42230-00e1-48e0-86ce-c04ab5cdf2f9_1456x648.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:648,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!U7kE!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7c42230-00e1-48e0-86ce-c04ab5cdf2f9_1456x648.png 424w, https://substackcdn.com/image/fetch/$s_!U7kE!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7c42230-00e1-48e0-86ce-c04ab5cdf2f9_1456x648.png 848w, https://substackcdn.com/image/fetch/$s_!U7kE!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7c42230-00e1-48e0-86ce-c04ab5cdf2f9_1456x648.png 1272w, https://substackcdn.com/image/fetch/$s_!U7kE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7c42230-00e1-48e0-86ce-c04ab5cdf2f9_1456x648.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://ai-2027.com/research/timelines-forecast#defining-a-superhuman-coder-sc">Source</a></figcaption></figure></div><p>I do not share this assumption that this will be so straightforward.</p><p>The speed up in doubling time in 24-25, from 23-24 can be explained by the shift to <em>training for long-horizonness </em>which wasn&#8217;t happening beforehand, as AI labs focused on scaling pre-training. (While the faster rate from 24-25 might indicate that we could expect long-horizon capabilities to grow faster, it doesn&#8217;t imply there is an exponential where the rate of the rate is constantly growing.) From the rumours I have heard, reinforcement learning is not generalising well to longer tasks. Therefore, a conceptual assumption that 1 month tasks don&#8217;t <em>feel </em>that different from 2 month-tasks for humans is a reasonably weak basis for such a consequential conclusion. We just don&#8217;t know. As the exponential extrapolation (purple line) shows, the time-horizonness is extremely sensitive to this assumption.</p><p>In spite of these methodological differences for extrapolation, I doubt that &#8220;time horizon&#8221; exchanges very closely with the kind of <em>usefulness </em>we&#8217;re actually looking for.</p><p>This aspect of the scenario is where I agree most closely with the authors: we should expect fast progress in software engineering agents. But I think expecting a superhuman coder in March 2027 isn&#8217;t &#8220;conservative&#8221;, it is aggressive. This kind of capability could arrive any time from 2027 to 2035, depending on <a href="https://inferencemagazine.substack.com/i/155018281/complete-automation-could-be-bottlenecked-by-ideation-and-research-taste">the kinds of holdout tasks that we get</a>. In any case, these &#8220;intermediate&#8221; agents would have some augmentative effect on research progress, and indeed, <em>already are.</em></p><h2>Using an automated researcher</h2><p>Where I strongly diverge from the authors, is <em>how useful </em>the superhuman coder will be.</p><p>Their method supposes once a superhuman coder has been created in March 2027, that it accelerates AI research by a factor of 5. This means a &#8220;superhuman AI researcher&#8221; is created in July 2027, which then accelerates AI research by a factor of 25. This leads to the creation of a &#8220;superintelligent AI researcher&#8221; in November 2027, which accelerates AI research by a factor of <em>250</em>. Which leads to the creation of artificial superintelligence in April 2028 which accelerates research by <em>2000</em> times. These multipliers are created by adding together estimates of different factor improvements <a href="https://www.getguesstimate.com/models/25630">viewable here</a>.</p><h3>What are the superhuman coders going to do?</h3><p>The scenario says that in March 2027, the AI lab is running 200,000 copies of the &#8220;superhuman coder&#8221; which is capable of <em>implementing experiments, </em>but not developing ideas at the level of the best human researchers. (By June, it is 250,000 copies)</p><p>The scenario has the lab using 6% of their compute for running these copies, and 25% for experiments. In their analysis, OpenAI has 20 million H100-equivalents in 2027, so 1.2 million H100e&#8217;s are used to run the agents and 5 million H100e&#8217;s go on experiments. This means there are only 25 H100s per &#8220;superhuman coder&#8221; in March.</p><p>This would be a suboptimal way to manage the compute budget. There is a tradeoff between running copies of the automated researcher and running more experiments, and the question has to be: <em>where are we most constrained?</em></p><p>The answer, I believe, is in experimental throughput.</p><p>AI research is an empirical field, where smaller scale results do not generalise to larger models. See this excerpt from Sholto Douglas on the Dwarkesh Podcast:</p><blockquote><p>&#8220;[Y]ou never actually know if the trend will hold. For certain architectures the trend has held really well. And for certain changes, it's held really well. But that isn't always the case. And things which can help at smaller scales can actually hurt at larger scales. You have to make guesses based on what the trend lines look like and based on your intuitive feeling of what&#8217;s actually something that's going to matter, particularly for those which help with the small scale.&#8221;</p></blockquote><p>I heard from one source that labs have to attempt experiments at 10 or 12 increments of scale before an architectural change might go into the next training run. These experiments can take multiple days or even weeks, depending on the size or compute allocation.</p><p>All labs are constrained by experimental compute at present. Conceptually, the lab ought to be limited by experimental compute, otherwise<em> </em>it would be constrained by something else (like ideas to try), which would be worse. And, right now, if the lab were to hire another research leader they would be forced to <a href="https://inferencemagazine.substack.com/i/155018281/the-amount-of-experimental-compute-that-ai-labs-have-places-limits-on-the-size-of-their-human-ai-researchers-teams">split their experimental budget by n+1 researchers</a>.</p><p>Aidan McLaughlin has <a href="https://x.com/aidan_mclau/status/1917271221827428548">said that</a> &#8220;every researcher is experimental compute constrained&#8221;, and Sholto has said that (while he was at DeepMind) &#8220;the Gemini program would probably be maybe five times faster with 10 times more compute or something like that&#8221;.</p><p>If this is the case, the critical determinant of research progress is <em>how widely and intuitively </em>you can search for new breakthroughs, and how many ideas you can try at a larger scale. c.f. Sholto again:</p><blockquote><p>&#8220;Many people have a long list of ideas that they want to try, but paring that down and shot calling, under very imperfect information, what are the right ideas to explore further is really hard.&#8221;</p></blockquote><p>This is certainly <em>somewhat </em>sensitive to having automated software engineers but I would dispute that it is sensitive by a factor of 5 and would suggest it is more sensitive to the overall size of the compute budget.</p><h3>How sensitive is research progress to superhuman coders?</h3><p>The work of an AI researcher has four main components: making hypotheses, designing experiments, supervising experiments, and analysing results.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!eVTg!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed285509-d6e0-40fc-97d0-08c713f009c8_1600x800.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!eVTg!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed285509-d6e0-40fc-97d0-08c713f009c8_1600x800.png 424w, https://substackcdn.com/image/fetch/$s_!eVTg!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed285509-d6e0-40fc-97d0-08c713f009c8_1600x800.png 848w, https://substackcdn.com/image/fetch/$s_!eVTg!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed285509-d6e0-40fc-97d0-08c713f009c8_1600x800.png 1272w, https://substackcdn.com/image/fetch/$s_!eVTg!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed285509-d6e0-40fc-97d0-08c713f009c8_1600x800.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!eVTg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed285509-d6e0-40fc-97d0-08c713f009c8_1600x800.png" width="1456" height="728" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ed285509-d6e0-40fc-97d0-08c713f009c8_1600x800.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:728,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!eVTg!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed285509-d6e0-40fc-97d0-08c713f009c8_1600x800.png 424w, https://substackcdn.com/image/fetch/$s_!eVTg!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed285509-d6e0-40fc-97d0-08c713f009c8_1600x800.png 848w, https://substackcdn.com/image/fetch/$s_!eVTg!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed285509-d6e0-40fc-97d0-08c713f009c8_1600x800.png 1272w, https://substackcdn.com/image/fetch/$s_!eVTg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed285509-d6e0-40fc-97d0-08c713f009c8_1600x800.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://ai-2027.com/research/takeoff-forecast#from-sc-to-sar">Source</a></figcaption></figure></div><p>Automating experimental design and supervising experiments would give researchers more time for studying results, reading others&#8217; work, and thinking about which experiments to run next but <em>whether output goes up</em> would depend on the differential quality of ideas they tried on their constrained compute or, if there is additional compute &#8216;freed up&#8217;, ideas they could have not have otherwise attempted.</p><p>However, if there were 100 researchers who shared a fixed compute budget, and they all gained 40% more time for generating ideas, automating implementation would <em>exacerbate the existing </em>compute constraint. Prioritising ideas, and where to search, again becomes the binding constraint.</p><p>The scenario highlights that cheap, fast superhuman coders could optimise compute usage by <em>flexibly prioritising</em> the highest priority work, <em>catching bugs</em>, <em>monitoring overnight experiments</em> (and restarting them if they break) and <em>running multiple independent variables </em>on a single experiment.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a> I think this kind of thing is useful in aggregate, and there are some forms of labour which are only usable if it is quick. For example, optimising the kernel (lower-level code) for a small scale test. But one has to ask: if the gains from these optimisations were so big, why didn&#8217;t the company hire a human researcher to do it? From conversations with researchers, the research infrastructure at some labs is pretty highly optimised already, for example, to manage the optimal allocation of experimental compute.</p><p>I leave it as an exercise for the reader to determine <em>how much speedup </em>they believe a superhuman coder would provide on overall research output. I do not think the organisations will be <em>5 times faster</em>. For me, the range is somewhere between, 20% faster and 3 times faster. Another way to frame whether the AI 2027 argument is convincing is, all else equal, which lab would you bet on: 1000 human researchers using 6 million H100s and 33.3k automated coders; or 1000 human researchers using 5 million H100s and 200k automated coders? I would opt for the former.</p><p>The same method is applied to estimating how much speedup comes from the &#8220;superhuman AI researcher&#8221; and the &#8220;superintelligent AI researcher&#8221;, of which even more copies are run. The <em>superhuman</em> researcher is <em>as good as </em>the labs&#8217; best human researcher and the <em>superintelligent researcher </em>is much better. In the authors&#8217; calculations, the thousands of copies of the superhuman researcher provides 25 times speedup, and the superintelligent researchers provide 250 times speedup. Readers will have to consider: to what extent do these allow us to bypass <em>experimental throughput constraints?</em> This can happen a variety of ways:</p><ul><li><p>Optimising computational resources.</p></li><li><p>Having better intuition for which small-scale results should be scaled-up.</p></li><li><p>Generating better small scale experiments, from improved research taste. (The authors&#8217; apply a multiplier to the superintelligent AI researcher of 1.5x to 5x for better ideas.)</p></li><li><p>&#8220;Thinking faster.&#8221; (For what it is worth, I don&#8217;t think that output is constrained by thinking speed nearly as much as, say, needing to wait for experimental results.)</p></li></ul><p>For me, the multiplier from each capability is much smaller than for the AI 2027 authors, and overall progress to be much less affected by the labour, than the compute available. When we reach AI systems which are much more capable of thinking of research ideas than humans, then progress could be extremely quick, but I think the initial expectations overestimate how soon this will be.</p><h3>Why are the labs putting so much compute towards R&amp;D?</h3><p>The R&amp;D budget of any company depends on the expectation of future profits.</p><p>The AI 2027 scenario imagines that AI labs will spend between 80% and 87% of their compute on R&amp;D in 2027, with the remaining 13-20% being spent on selling models to customers. This depends on aggressive revenue assumptions. In Q2 2027, the scenario predicts that the leading lab would be doing $120 billion in revenue (and servicing that with just 1.71 million H100-equivalents!).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a> By contrast, OpenAI expects to hit $125 billion revenue <em>in 2029</em>. In AI 2027&#8217;s scenario, revenue is $8 trillion and&#8230;</p><blockquote><p>&#8220;Humans realize that they are obsolete. A few niche industries still trade with the robot economy, supplying goods where the humans can still add value.30 Everyone else either performs a charade of doing their job&#8212;leaders still leading, managers still managing&#8212;or relaxes and collects an incredibly luxurious universal basic income.&#8221;</p></blockquote><p>This is not a reasonable, especially not a &#8220;conservative&#8221;, assumption for automation.</p><p>The AI 2027 prediction of $100 billion in 2027 is based on very flimsy <a href="https://futuresearch.ai/openbrain-revenue">analysis</a> by a third party. This group proposes two methodologies: first, they extrapolate how quickly companies are reaching $100 billion in revenue and naively extrapolate this time to OpenAI.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-A0-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b6f5307-20e9-4b15-84de-bc18f1f730f4_1346x906.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-A0-!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b6f5307-20e9-4b15-84de-bc18f1f730f4_1346x906.png 424w, https://substackcdn.com/image/fetch/$s_!-A0-!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b6f5307-20e9-4b15-84de-bc18f1f730f4_1346x906.png 848w, https://substackcdn.com/image/fetch/$s_!-A0-!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b6f5307-20e9-4b15-84de-bc18f1f730f4_1346x906.png 1272w, https://substackcdn.com/image/fetch/$s_!-A0-!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b6f5307-20e9-4b15-84de-bc18f1f730f4_1346x906.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-A0-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b6f5307-20e9-4b15-84de-bc18f1f730f4_1346x906.png" width="1346" height="906" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5b6f5307-20e9-4b15-84de-bc18f1f730f4_1346x906.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:906,&quot;width&quot;:1346,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!-A0-!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b6f5307-20e9-4b15-84de-bc18f1f730f4_1346x906.png 424w, https://substackcdn.com/image/fetch/$s_!-A0-!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b6f5307-20e9-4b15-84de-bc18f1f730f4_1346x906.png 848w, https://substackcdn.com/image/fetch/$s_!-A0-!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b6f5307-20e9-4b15-84de-bc18f1f730f4_1346x906.png 1272w, https://substackcdn.com/image/fetch/$s_!-A0-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b6f5307-20e9-4b15-84de-bc18f1f730f4_1346x906.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Second, the model hinges on &#8220;replacement workers&#8221;. They list 300,000 customer service reps, 180,000 knowledge workers, 90,000 software engineering agents and more than 90,000 R&amp;D researchers (which customers pay $20k per month for). This pace of automation would be unprecedented. Most industrial revolutions have produced 0.5-1% uplift in total factor productivity annually for decades. I expect AI to produce greater and faster uplift to productivity than this, there are still bottlenecks to deployment which I wrote about with my coauthor <a href="https://inferencemagazine.substack.com/i/155018281/cognitive-labour-will-be-automated-before-physical-labour-and-could-be-automated-much-more-quickly-than-previous-technological-revolutions">here</a>.</p><p>To contextualise the claim that OpenAI might have $100 billion in revenue by 2027, commercial Microsoft Office 365 was <a href="https://www.microsoft.com/en-us/investor/earnings/fy-2024-q4/productivity-and-business-processes-performance?utm_source=chatgpt.com">slightly under $50 billion in revenue in 2024</a>.</p><p>If we model the total cost of ownership for an accelerator at $70k for four years, the compute budget AI 2027 proposes has $39.15 billion for R&amp;D compute in Q4 2027. The investors and labs would have to answer where the incremental gross profit is going to come from, to sustain that rate of investment. (Especially difficult when <a href="https://inferencemagazine.substack.com/i/155018281/the-economically-useful-life-of-a-model-is-short">each model tends to depreciate so quickly</a>.)</p><p>Overall research output is most sensitive to <em>growth in R&amp;D compute, </em>because of its effects on <em>experimental throughput</em>. But the authors&#8217; expectations for R&amp;D compute budgets are downstream of ungrounded expectations for automation and revenue. With more grounded expectations for automation, R&amp;D budgets would be lower, and so research output would be less, so capabilities progress more slowly, so automation happens at a more reasonable pace.</p><h2>AI &#8220;arms race&#8221;</h2><p>The idea of an AI arms race hinges on two assumptions:</p><ol><li><p>That very small differences in capabilities &#8220;pre-takeoff&#8221; (automated research) confer very large differences in future capabilities because of multiplier effects (like 2500 times in AI 2027).</p></li><li><p>That differential capabilities will confer <em>decisive strategic advantage </em>on one country, over all others.</p></li></ol><p>For reasons discussed earlier, I&#8217;m unsure whether &#8220;multiplier effects&#8221; from automated researchers will get as large as AI 2027 expects, until much later.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a> Second, I don&#8217;t think it is yet clear that AI would confer decisive strategic advantage. <em>Perhaps it could</em> &#8212; and the perception that it might could be enough to set off an arms race &#8212; but this isn&#8217;t immediately obvious to me. New AI weapons will interact with the existing balance of power and deterrence framework.</p><p>Indeed, in all future scenarios, countries will compete forthrightly to have better AI and better deployment, but I don&#8217;t think it is certain to take on a &#8220;do-or-die&#8221; character for nation states. The narrative takes both aspects to be true, and assumes that leaders will be <em>extremely</em> cavalier about the strategic balance. It says:</p><blockquote><p>In cooperation with the military, [Agent-5] could help with defense R&amp;D, conduct untraceable cyberattacks on China, and win a decisive victory in the arms race.<br><br>The Oversight Committee is jubilant. Now is the decisive moment to beat China!</p><p>&#8230;<br><br>The American public mostly supports going to the bargaining table. &#8220;Why stop when we are winning?&#8221; says OpenBrain leadership to the President. He nods. The race continues.</p><p>&#8230;<br><br>After consulting with his advisors and the Oversight Committee, the President opts for the &#8220;We win, they lose&#8221; strategy. Perhaps China won&#8217;t go to war after all, and if they do, a deal can probably be made before it goes nuclear.</p></blockquote><p>This is <strong>extremely unrealistic</strong> and does not reflect how Great Powers think about strategic stability at all.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a> Both sides are interested in maintaining balance and moving competition out of the sphere of nuclear brinkmanship and arms stockpiling, into, say, economic adoption and diffusion. For both sides, the idea that deterrence could be undermined is extremely scary for what should be obvious reasons. So countries might be threatened not just by having decisive advantage but by the perception they might have it. Nobody would be &#8220;jubilant&#8221;. Everyone would understand how destabilising this could be; and both sides have interest, to the extent possible, in verifying the capabilities of the other and having their own verified.</p><p>When we first developed nuclear weapons, Bertrand Russell wrote in &#8220;<a href="https://scispace.com/pdf/the-atomic-bomb-and-the-prevention-of-war-1unnn2w25g.pdf">The Atomic Bomb and the Prevention of War</a>&#8221; that the United States should threaten and/or start another World War before the Soviet Union made nuclear weapons, and create a world government to prevent anyone else from developing them. With hindsight, we see this would have caused enormous suffering, cost the Free World its moral authority following World War Two, and exposed the risk of world government. Similar narratives for AI only increase the risk of bad outcomes, as humanity builds technology with uncertain strategic effects. Saffron Huang, a researcher at Anthropic, <a href="https://x.com/saffronhuang/status/1907863453009867183">said of the scenario</a>:</p><blockquote><p>They say they don't want this scenario to come to pass, but their actions---trying to make scary outcomes seem unavoidable, burying critical assumptions, burying leverage points for action---make it more likely to come to pass.</p></blockquote><h3>Nation states (not the US and China)</h3><p>The scenario focuses on the US-China relationship, naturally, but casts all other nations as background extras. In the scenario, in May 2027, &#8220;America&#8217;s foreign allies are out of the loop&#8221; including UK AISI; in October&#8230;</p><blockquote><p>&#8220;Foreign allies are outraged to realize that they&#8217;ve been carefully placated with glimpses of obsolete models. European leaders publicly accuse the US of &#8220;creating rogue AGI&#8221; and hold summits demanding a pause, with India, Israel, Russia, and China all joining in.&#8221;</p></blockquote><p>This is a Bay Area bubble view. Other countries will not be <em>this irrelevant</em>. Because I do not share their expectations for progress, it is a little difficult to comment directly on the scenario.  In general, I expect that frontier capabilities will be more public because the labs have to productise models to pay for R&amp;D. So everyone will have a better sense for AI progress. In the last year, it seems AI labs have accelerated their product development cycles &#8212; <a href="https://x.com/jxmnop/status/1919057585581478229?s=46">&#8220;normalising&#8221; into big tech companies</a> &#8212; rather than acting like AI R&amp;D and internal deployment is <em>by far </em>the most important thing.</p><p>If countries <em>knew</em> they were in the dark about AI progress, this would be concerning and destabilising. Therefore it would be unlikely to help global security and therefore improbable, though not impossible, that Great Powers would try to make their AI development secret. One has to consider in much more depth how the strategic balance changes for <em>all countries in the world, </em>and what they would do to prevent being undermined.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-10" href="#footnote-10" target="_self">10</a></p><h2>Conclusion</h2><p>This scenario is not <em>so</em> scary to me because its bad outcomes depended on leaders taking irresponsible actions. I would think it was much more dangerous if everyone had behaved defensibly and things still went bad.</p><p>This is revealing. I find it quite difficult to specify how a responsible actor (either within political or lab leadership positions) should be acting. For any action one can recommend, there are sensible counter-arguments why it is less obvious. Pausing AI research <em>now</em> is not compelling. Instituting global governance and <a href="https://inferencemagazine.substack.com/p/the-uphill-battle-to-mitigate-the">any national regulation</a> also have counterarguments. From a distance, <strong>responsible action and irresponsible action look quite similar. </strong>Replay nuclear politics from 1945 onwards. Is there anything that can be said <em>generally </em>about what constitutes responsible action at each step? (Sure, I think we can agree that Nixon shouldn&#8217;t have ordered retaliatory strikes whilst he was drunk.) That period of history was very dangerous but there wasn&#8217;t <em>much </em>which could have made it less so.</p><p>What scares me about AI progress is that things might happen too quickly and this does not give us the time to respond. To give credit where it is due, the authors have compellingly raised the salience of <em>what this could be like </em>for those unfamiliar with the field, and <em>what could be at stake </em>for those in the room. While at times, the analysis gets caught in the Bay Area&#8217;s eschatological dialectic, the essence is defendable: AI progress is going fast, and can move faster still. Perhaps so fast we cannot even process it.</p><p></p><p>Come what may, we&#8217;ll have to do our best.</p><div><hr></div><blockquote><p>Given a total lack of independent intellectual steering power and no desire to spend thirty years building an independent knowledge base of Near Eastern history, I choose to just accept the ideas of the prestigious people with professorships in Archaeology, rather than those of the universally reviled crackpots who write books about <a href="http://en.wikipedia.org/wiki/Worlds_in_Collision">Venus being a comet</a>.</p><p>You could consider this a form of epistemic learned helplessness, where I know any attempt to evaluate the arguments is just going to be a bad idea so I don&#8217;t even try. If you have a good argument that the Early Bronze Age worked completely differently from the way mainstream historians believe, I just don&#8217;t want to hear about it. If you insist on telling me anyway, I will nod, say that your argument makes complete sense, and then totally refuse to change my mind or admit even the slightest possibility that you might be right.<br><br>&#8212; Scott Alexander, <a href="https://slatestarcodex.com/2019/06/03/repost-epistemic-learned-helplessness/">Epistemic Learned Helplessness</a></p></blockquote><p></p><p></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>&#8220;When reading the works of an important thinker, look first for the apparent absurdities in the text and ask yourself how a sensible person could have written them. When you find an answer, I continue, when those passages make sense, then you may find that more central passages, ones you previously thought you understood, have changed their meaning.&#8221; &#8212; Thomas Kuhn, The Essential Tension (1977), xii.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Extended definition: &#8220;An AI system for which the company could run with 5% of their compute budget 30x as many agents as they have human research engineers, each of which is on average accomplishing coding tasks involved in AI research (e.g. experiment implementation but not ideation/prioritization) at 30x the speed (i.e. the tasks take them 30x less time, not necessarily that they write or &#8220;think&#8221; at 30x the speed of humans) of the company&#8217;s best engineer. This includes being able to accomplish tasks that are in any human researchers&#8217; area of expertise. Nikola and Eli estimate that the first SC will have at least 50th percentile frontier AI researcher &#8220;research taste&#8221; as well, but that isn&#8217;t required in the definition.&#8221;</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>All SWE-bench-verified scores are pass@1</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>&#8220;I would guess that sometime in the next 12 to 18 months, we'll reach the point where most of the code that's going toward these efforts is written by AI. And I don't mean autocomplete. Today you have good autocomplete. You start writing something and it can complete a section of code. I'm talking more like: you give it a goal, it can run tests, it can improve things, it can find issues, it writes higher quality code than the average very good person on the team already.&#8221; &#8211; Zuckerberg on Dwarkesh Podcast</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>I don&#8217;t expect much to come from running multiple IVs. Most labs seem to be trying to increase empiricism and decrease intuition for how to run their research.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>This was calculated from looking at the compute budget table and the predictions of revenue on the main scenario.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>From the piece; &#8220;applying this to OpenAI would indicate $100B revenue by mid-2027, which is consistent with our simple exponential growth model.&#8221;</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p> i.e. when AI systems far surpass human ability at picking ideas, and even then&#8230;</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p>See works like &#8220;The Strategy of Conflict&#8221; by Thomas Schelling</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-10" href="#footnote-anchor-10" class="footnote-number" contenteditable="false" target="_self">10</a><div class="footnote-content"><p>There is very little work on &#8220;AGI and the strategic balance in [Eastern Europe / Israel-Palestine conflict / the broader Middle East / India-Pakistan / South East Asia / South America]&#8221;.</p></div></div>]]></content:encoded></item><item><title><![CDATA[“And then we get the robots”]]></title><description><![CDATA[Progress in robotics isn't just an intelligence problem.]]></description><link>https://inferencemagazine.substack.com/p/and-then-we-get-the-robots</link><guid isPermaLink="false">https://inferencemagazine.substack.com/p/and-then-we-get-the-robots</guid><dc:creator><![CDATA[Jack Wiseman]]></dc:creator><pubDate>Wed, 30 Apr 2025 15:59:53 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c61df20-b545-4c7b-9acb-75940496383f_1010x1010.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Most of the narrative accounts of the intelligence explosion predict <em>very fast</em> robotics progress will follow the invention of powerful AI. (See <a href="https://ai-2027.com/slowdown#slowdown-2028-02-29">here</a>, <a href="https://situational-awareness.ai/from-agi-to-superintelligence/#The_power_of_superintelligence">here</a>, and <a href="https://www.forethought.org/research/preparing-for-the-intelligence-explosion#the-industrial-explosion">here</a>.) But none of these predictions are specific about <em>exactly what </em>needs to be solved. They gesture at data bottlenecks but ultimately abstract away the challenges to robotic progress and imagine that superintelligent AI could &#8220;solve robotics&#8221;. If it were possible to do this, it could well lead to much more dramatic labour automation and explosive economic growth. So we need a better understanding of <em>how sensitive </em>robotics progress is to AI progress. </p><p>There are a few Silicon Valley companies aiming to build humanoid robots at present. Their robots are made from electromechanical actuators, gearboxes, and cameras (though some also have LiDAR or proprioceptive sensors embedded). While these robots can produce impressive demonstrations, they cannot behave reliably in general environments. There are not a large or diverse enough set of training data for them to learn general behaviour. Language models have the luxury of an enormous dataset to make into a prediction task, where robots do not. Reinforcement learning is much cheaper than text than robots. &#8220;Generating a trajectory&#8221; means trying to solve a maths problem on a digital trackpad, not trying to drive a car. Failing is cheaper too. Language model hallucinations are quaint, self-driving car hallucinations&#8230;aren&#8217;t.</p><p>There are three ways to solve the data problem. The first is simply to <strong>gather more data from robots trying to solve problems</strong>. Google tried to solve this by building an &#8220;arm farm&#8221; and having these robots attempt to solve problems around the clock, but this was closed. Others have had humans complete tasks using robot grippers or had humans teleoperate robots to gather data. Aside from increasing the amount of data, <strong>researchers can improve the training procedure, so the robot &#8220;learns more&#8221; per example</strong>. The chart below shows how, over time, equivalent performance on an image recognition task required less data and computational resources.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Nz4W!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e3af632-7567-4412-8efe-3c84465f0a2a_1600x1076.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Nz4W!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e3af632-7567-4412-8efe-3c84465f0a2a_1600x1076.png 424w, https://substackcdn.com/image/fetch/$s_!Nz4W!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e3af632-7567-4412-8efe-3c84465f0a2a_1600x1076.png 848w, https://substackcdn.com/image/fetch/$s_!Nz4W!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e3af632-7567-4412-8efe-3c84465f0a2a_1600x1076.png 1272w, https://substackcdn.com/image/fetch/$s_!Nz4W!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e3af632-7567-4412-8efe-3c84465f0a2a_1600x1076.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Nz4W!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e3af632-7567-4412-8efe-3c84465f0a2a_1600x1076.png" width="1456" height="979" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5e3af632-7567-4412-8efe-3c84465f0a2a_1600x1076.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:979,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Nz4W!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e3af632-7567-4412-8efe-3c84465f0a2a_1600x1076.png 424w, https://substackcdn.com/image/fetch/$s_!Nz4W!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e3af632-7567-4412-8efe-3c84465f0a2a_1600x1076.png 848w, https://substackcdn.com/image/fetch/$s_!Nz4W!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e3af632-7567-4412-8efe-3c84465f0a2a_1600x1076.png 1272w, https://substackcdn.com/image/fetch/$s_!Nz4W!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e3af632-7567-4412-8efe-3c84465f0a2a_1600x1076.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://epoch.ai/blog/revisiting-algorithmic-progress">Source</a></figcaption></figure></div><p>Otherwise, it could become possible in the future to <strong>create simulated environments for the robots to train in</strong>. There are academic examples of researchers teaching robots to perform <a href="https://techxplore.com/news/2024-04-sim-real-robots-simple-tasks.html?utm_source=chatgpt.com">very simple tasks</a> using synthetic data, but it is difficult to create realistic environments for robots to learn more complicated tasks. Google DeepMind and OpenAI are trying to create physical world models. DeepMind&#8217;s model, <a href="https://deepmind.google/discover/blog/genie-2-a-large-scale-foundation-world-model/">Genie 2,</a> was trained on video game data to turn an image into a consistent 3-d world for up to a minute. There&#8217;s a large amount of compute scaling and data which could improve these models, but of course, it's uncertain how far this would go: it is much easier to imagine these very large models helping with planning tasks and coarse movements than fine motor control.</p><h3>How could automated AI progress help with the data problem?</h3><p>Big jumps in AI capabilities could accelerate the solution to the data problem in a couple of ways. Most directly, automated researchers could design better training algorithms, new model architectures and so on. However, we cannot assume that the rate of algorithmic improvement in AI researchers can generalise to robots. Researchers have to test whether their changes have worked, and testing whether an automated AI researcher is better can be done entirely within a computer whereas knowing if the automated robotics researcher can design better algorithms requires real-world (expensive) feedback. This could, however, be slightly mitigated. Automated researchers could help to design better simulated environments for the robots to train in. All of this automated software progress is <a href="https://inferencemagazine.substack.com/i/155018281/to-what-extent-can-ai-labs-maintain-an-experimental-compute-budget">subject to the same constraints</a> as automated AI research.</p><p>But <em>even if </em>superintelligent AI could generate entirely realistic simulations and find the optimal learning algorithm or model architecture for the current set of hardware, robots wouldn&#8217;t be able to complete all the tasks humans can. We would run into hard limits. From a recent <a href="https://www.construction-physics.com/p/robot-dexterity-still-seems-hard">Construction Physics article</a> on robot dexterity:</p><blockquote><p>Human hands are very strong while being capable of complex and precise motions, and it&#8217;s difficult to match this with a robot hand. Robot hands are often surprisingly weak. An <a href="https://outlift.com/how-much-can-the-average-man-lift/#3-how-much-can-the-average-man-deadlift">average man</a> has enough grip strength to lift 40 kg or more off the ground (20 kg in each hand), and a strong man can lift upwards of 100 kg. By contrast, NASA&#8217;s <a href="https://ntrs.nasa.gov/api/citations/20110023122/downloads/20110023122.pdf">Robonaut 2 hand</a> had a payload capacity of 9 kilograms, and the <a href="https://www.shadowrobot.com/dexterous-hand-series/">Shadow dexterous hand</a> (billed as the &#8220;most advanced 5-fingered robotic hand in the world&#8221;) has a payload capacity of just 4 kilograms.</p><p>More importantly, human hands are extremely sensitive, and capable of providing a lot of tactile feedback to help guide our actions. A human hand has around <a href="https://www.ncbi.nlm.nih.gov/books/NBK279362/#:~:text=Our%20hands%20also%20have%20very,nerve%20endings%20in%20the%20palm.">17,000 touch receptors</a>, and is sensitive enough to discriminate between textures that differ by <a href="https://pmc.ncbi.nlm.nih.gov/articles/PMC3771396/">mere nanometers</a>. Robot hands <a href="https://www.youtube.com/watch?v=rj4W_S48b0U">are</a> <a href="https://x.com/HumanoidRTech/status/1912780655676489895">getting better</a>, but still don&#8217;t appear to be close to what a human hand can do. <a href="https://www.youtube.com/watch?v=AXcZGkKR2og">This robot hand</a>, for instance, boasts &#8220;17 tactile sensors,&#8221; and <a href="https://www.youtube.com/watch?v=0rwYOa7pJCs">this one from Unitree</a> has 94.</p></blockquote><p>Optimising the current hardware would unlock some economically-useful tasks but not the &#8220;100% of human tasks&#8221; that predictions of the intelligence explosion would require. Getting closer to all physical labour being automated would require a leap forward in hardware progress.</p><h3>How sensitive is robotic hardware R&amp;D to intelligence?</h3><p>ARIA has a research programme dedicated to <a href="https://www.aria.org.uk/opportunity-spaces/smarter-robot-bodies/robot-dexterity/">improving robotic hardware</a> which can provide a guide to the kinds of step-changing inventions necessary.</p><p>The next step for robotic sensing is to improve tactile sensing. One way to&nbsp;think about sensing is that it has been free riding on improvements in cameras and so vision capabilities far outpace other modalities. The ARIA programme has funded three tactile sensing projects. One group has developed a new material that conducts electricity in proportion to the force being applied to it, and will combine this with directional strain and temperature sensors into a single e-skin. Another group has developed a material which is continuously conducting and uses changes to the voltage to identify force on the surface. A final group is using changes in magnetic fields to allow for very granular sensing.</p><p>Electromechanical actuators are limited in a number of ways. They have low torque density, limited bandwidth (reaction time), high inertia, and they scale down very poorly. In the limit, a DC motor becomes a heat machine. The ARIA programme is funding alternative muscles:</p><ul><li><p>For very fine control (e.g. finger joints or semiconductor assembly) one group has developed a material layered with liquid droplets and electrodes between the layers. When an electric current is applied to the electrodes, it causes the liquid droplets to compress, creating a &#8220;gripping&#8221; effect when the material is stacked in layers.</p></li><li><p>A group is developing a braided material for pneumatic (air or liquid) muscles that can channel the radial force during &#8220;contraction&#8221; to make control more precise.</p></li><li><p>A group is making a new material geometry for an elastomer. The material contracts like a muscle when current runs through it but this currently requires very high voltage and so the project is aiming to bring this down.</p></li><li><p>A related project is developing a muscle mimic which moves fluid around in a soft pouch.</p></li><li><p>Finally there is a group working on synthetic muscle fibres.</p></li></ul><p>There are three projects which aim to reduce the number of gearboxes a robot would need:</p><ul><li><p>One project is replacing the gearbox with an arrangement of magnets at different polarities to control rotary motion.</p></li><li><p>Another project is miniaturising an existing actuator that has pairs of magnets controlling linear motion.</p></li><li><p>A third project is developing clutches which could route power from a single gearbox to reduce the need for every joint to have a gearbox.</p></li></ul><p>From the outside, very powerful AI would be very useful for aspects of the R&amp;D process but wouldn&#8217;t &#8220;solve&#8221; the problem end-to-end. Many of the processes require materials R&amp;D and AI is very useful for discovery and modelling behaviour. But these processes ultimately depend on real-world experimental data to train the models and  to refine the search. Similarly, for the sensing projects, very capable AI could suggest topologies for spacing the sensors, to account for wiring complexity, the flexibility of the material, cost, and so on; but computer simulations would be an imperfect substitute for seeing how robust the material was to a month&#8217;s intense use. Very good AI should minimise, but not totally eliminate, iteration cycles. Prototyping and manufacturing for real-world experimentation becomes binding.</p><p>Were robotics progress going to happen very quickly, all of the tasks involved in hardware R&amp;D would need to become &#8220;intelligence problems&#8221;. The crux here is to what degree do humans need to be involved in the iteration cycle: how good can the physics simulations get,<em> </em>that hardware design happens entirely in silico? Can the scientific agents figure out, say, the optimal arrangement of magnets and their strengths to control a robotic &#8220;shoulder&#8221; <em>while also </em>trading off weight, manufacturability, durability and so on? Are <em>all</em> of the questions answered by simulation?</p><p>The idea that a technological singularity will occur after we automate AI research abstracts away these practical bottlenecks in the R&amp;D process. <em>AI is going to change everything</em>, but it won&#8217;t be overnight.</p>]]></content:encoded></item><item><title><![CDATA[The Parrot is Dead]]></title><description><![CDATA[And we should deal with it.]]></description><link>https://inferencemagazine.substack.com/p/the-parrot-is-dead</link><guid isPermaLink="false">https://inferencemagazine.substack.com/p/the-parrot-is-dead</guid><dc:creator><![CDATA[Jack Wiseman]]></dc:creator><pubDate>Fri, 11 Apr 2025 01:55:13 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/151e0519-4d15-4dfc-9933-10e7013262d0_2838x1882.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!DhFA!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fddd0a233-fe71-4265-8f09-2b8a7d0e81d1_304x236.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!DhFA!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fddd0a233-fe71-4265-8f09-2b8a7d0e81d1_304x236.png 424w, https://substackcdn.com/image/fetch/$s_!DhFA!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fddd0a233-fe71-4265-8f09-2b8a7d0e81d1_304x236.png 848w, https://substackcdn.com/image/fetch/$s_!DhFA!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fddd0a233-fe71-4265-8f09-2b8a7d0e81d1_304x236.png 1272w, https://substackcdn.com/image/fetch/$s_!DhFA!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fddd0a233-fe71-4265-8f09-2b8a7d0e81d1_304x236.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!DhFA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fddd0a233-fe71-4265-8f09-2b8a7d0e81d1_304x236.png" width="612" height="475.10526315789474" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ddd0a233-fe71-4265-8f09-2b8a7d0e81d1_304x236.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;normal&quot;,&quot;height&quot;:236,&quot;width&quot;:304,&quot;resizeWidth&quot;:612,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!DhFA!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fddd0a233-fe71-4265-8f09-2b8a7d0e81d1_304x236.png 424w, https://substackcdn.com/image/fetch/$s_!DhFA!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fddd0a233-fe71-4265-8f09-2b8a7d0e81d1_304x236.png 848w, https://substackcdn.com/image/fetch/$s_!DhFA!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fddd0a233-fe71-4265-8f09-2b8a7d0e81d1_304x236.png 1272w, https://substackcdn.com/image/fetch/$s_!DhFA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fddd0a233-fe71-4265-8f09-2b8a7d0e81d1_304x236.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a><figcaption class="image-caption"><a href="https://en.wikipedia.org/wiki/Dead_Parrot_sketch">Source</a></figcaption></figure></div><blockquote><p>This parrot is no more! He has ceased to be! He's expired and gone to meet his maker! He's a stiff! Bereft of life, he rests in peace! If you hadn't nailed him to the perch, he'd be pushing up the daisies! His metabolic processes are now history! He's off the twig! He's kicked the bucket, he's shuffled off his mortal coil, run down the curtain and joined the bleedin' choir invisible! THIS IS AN EX-PARROT!<br><br>&#8212; <a href="https://www.youtube.com/watch?v=4vuW6tQ0218">Monty Python, Series 1, Episode 8</a></p></blockquote><p>For a while, some people dismissed language models as &#8220;stochastic parrots&#8221;. They said models could just memorise statistical patterns, which they would regurgitate back to users. A model was a <em>simulacrum </em>of intelligence: it would mimic patterns of intelligent thought, but never go beyond the data it had seen in training.</p><p>The problem with this theory, is that, alas, it isn&#8217;t true.</p><p>And fortunately for our purposes, <em>exactly how</em> the parrot &#8216;ceased to be&#8217; is a good hook for explaining what&#8217;s going on inside language models.</p><div><hr></div><p>If a language model was just a stochastic parrot, when we looked inside to see what was going on, we&#8217;d basically find a lookup table. The model would be <em>embedding </em>its input sequence (read: turning a string of words into a matrix) and running a search for the most similar pattern in its training data and copying this. But it doesn&#8217;t look like this. As we delve into the models, we find circuits. These are general algorithms that the model has made to solve classes of problems.</p><p>These circuits aren&#8217;t &#8216;laid out&#8217; like how an electrician would wire a house or a programmer would write a program. They are more like an unholy tangle of wires. This is&#8212;counterintuitively&#8212;desirable! The goal of a model is to most <em>efficiently </em>represent all of the information and to <em>generalise </em>to solve problems. If the researchers had laid out how the model should achieve this, it would be more of a hindrance than a help. We want to <a href="https://inferencemagazine.substack.com/p/on-o1">&#8220;let the compute figure it out&#8221;</a>.</p><p>What does this mean in practice? A model is a stack of layers that contain a sequence of mathematical operations. The researchers control the &#8216;settings&#8217;, like the number of layers in the model and the learning policy. The model learns its &#8216;weights&#8217; (read: values for the mathematical operations). Through the complex interaction between weights, the models learn circuits<em>. </em>So the circuits are controlling how information flows through the layers.</p><p>This means circuits aren&#8217;t easy to spot. The first time they were seen in language models was December 2021, when Anthropic <a href="https://transformer-circuits.pub/2021/framework/index.html#induction-heads">released a paper</a> pointing to &#8216;induction heads&#8217;. These were a kind of attention head that could notice patterns in the input sequence and so, on the fly, the model could realise that it might need to recreate this pattern later. For the <em>general </em>class of problems&#8212;&#8220;recreate patterns found in the input&#8221;&#8212;this circuit could be reused for other patterns the model hadn&#8217;t seen in training. This is clearly more than the rote memorisation which AI sceptics had said language models were doing!</p><p>Until recently, the circuits that had been identified were limited to &#8220;toy-sized&#8221; models and often &#8220;algorithmic&#8221; tasks, not the full complex behaviour of large models that we care about. (For example, this induction heads paper had been on a two layer model.) This changed with <a href="https://transformer-circuits.pub/2024/scaling-monosemanticity/">a May 2024 paper</a> from Anthropic, which developed a technique to elicit representations (read: sub-units of circuits) from much larger models. You might&#8217;ve seen this already, because they used their elicitation techniques to find a &#8220;Golden Gate Bridge&#8221; feature and develop a demo which had this always turned on. Whenever you gave this version of Claude a prompt, it would find a way to steer its answer towards the Golden Gate Bridge <em>regardless </em>of the original topic.</p><p>Their <a href="https://transformer-circuits.pub/2025/attribution-graphs/biology.html">latest</a> <a href="https://transformer-circuits.pub/2025/attribution-graphs/methods.html">papers</a>, from a couple of weeks ago, are extending this work to show how the model combines and relates these internal representations of features like &#8220;Golden Gate Bridge&#8221; to form circuits. The most important bit of this is that they elicit circuits for complex behaviours in large models. This proves that even in more complex situations than &#8220;patterns from the input&#8221;, the model isn&#8217;t just a giant lookup it is doing <em>serious </em>computation. The kind which generalises.</p><h3><strong>What did they do?</strong></h3><p>The researchers built a new tool for seeing which features are active at every layer of the model. This means we can see <em>how </em>and <em>in what order </em>the model considers different bits of information. This is called a &#8220;cross-layer transcoder&#8221;. You can think of it as a string of lights attached to each layer. When a light is on, it shows which feature is activated. The researchers use these lights to assemble &#8220;attribution graphs&#8221;.</p><p>The most interesting application, in my view, was to how the model is generating rhyming couplets. There are two ways you might imagine this happening: it either improvises word-by-word as it goes or it plans ahead. The researchers found the latter&#8212;once the model had generated the first line, it would &#8220;look ahead&#8221; to the end of the second line.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!QCdZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b38b06d-eab0-487c-a1db-1bff299c9429_1528x1412.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!QCdZ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b38b06d-eab0-487c-a1db-1bff299c9429_1528x1412.png 424w, https://substackcdn.com/image/fetch/$s_!QCdZ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b38b06d-eab0-487c-a1db-1bff299c9429_1528x1412.png 848w, https://substackcdn.com/image/fetch/$s_!QCdZ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b38b06d-eab0-487c-a1db-1bff299c9429_1528x1412.png 1272w, https://substackcdn.com/image/fetch/$s_!QCdZ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b38b06d-eab0-487c-a1db-1bff299c9429_1528x1412.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!QCdZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b38b06d-eab0-487c-a1db-1bff299c9429_1528x1412.png" width="667" height="616.1504120879121" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8b38b06d-eab0-487c-a1db-1bff299c9429_1528x1412.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1345,&quot;width&quot;:1456,&quot;resizeWidth&quot;:667,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!QCdZ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b38b06d-eab0-487c-a1db-1bff299c9429_1528x1412.png 424w, https://substackcdn.com/image/fetch/$s_!QCdZ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b38b06d-eab0-487c-a1db-1bff299c9429_1528x1412.png 848w, https://substackcdn.com/image/fetch/$s_!QCdZ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b38b06d-eab0-487c-a1db-1bff299c9429_1528x1412.png 1272w, https://substackcdn.com/image/fetch/$s_!QCdZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b38b06d-eab0-487c-a1db-1bff299c9429_1528x1412.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>At the end of the first line (after the word &#8220;it&#8221;) the model would activate features which correspond to potential rhymes. In this case, &#8220;habit&#8221; or &#8220;rabbit&#8221; rhyme with &#8220;grab it&#8221;. The researchers perturbed the model activations to confirm the decision was actually happening at this point.</p><p>Next&#8212;and perhaps even more surprisingly&#8212;the model would use the final word it planned (&#8220;rabbit&#8221;) to plan the intermediate words. A group of &#8220;comparison features&#8221; were activated by the final word to analyse potential intermediates, so it was also <em>looking backwards </em>to create the structure of the line.</p><p>This kind of circuitry&#8212;to plan forwards and back&#8212;was learned by the model without explicit instruction; it just emerged from trying to predict the next word in other poems.</p><p>The challenge for research is that generalisation like this depends on seeing quite a lot of examples. Fortunately, there is a lot of poetry. The hypothesis we have for circuit formation is that initially, models are memorising a lookup table, but seeing enough examples causes the model to "privilege" the circuit rather than the table. The model realises its performance will be better from swapping to the general approach and thus, it seems like a circuit &#8220;snaps&#8221; into place. The paper includes some other examples where the model failed to make these generalisations because it hadn&#8217;t seen enough examples in those domains.</p><h3><strong>Where does this leave us?</strong></h3><p>The examples of generalisation are enough to prove the models are not stochastic parrots. But there are clearly limits to their circuitry. The more interesting version of the parrot debate is whether language models can generalise beyond circuits for &#8216;low level&#8217; tasks they&#8217;ve seen in their training data? Have they learned the ability to solve entirely <em>new </em>problems from seeing a large and diverse enough set of problems during training. While they&#8217;ve learned the muscle for generating rhyming couplets in novel ways, could they have come up with the concept in the first place?</p><p>This is the essence of Francois Chollet&#8217;s critique of language models (and his motivation for creating the ARC-AGI benchmark): <em>while they can learn circuits</em>, the real test is whether they can generate <em>new circuits on the fly </em>to solve unfamiliar problems. From <a href="https://www.dwarkesh.com/p/francois-chollet">a Dwarkesh interview</a>, Chollet says:</p><blockquote><p>LLMs are very good at memorizing small static programs. They've got this sort of bank of solution programs. When you give them a new puzzle, they can just fetch the appropriate program and apply it. It looks like reasoning but it's not really doing any sort of on-the-fly program synthesis. All it's doing is program fetching.</p><p>You can actually solve all these benchmarks with memorization. If you look at the models and what you're scaling up here, they are big parametric curves fitted to a data distribution. They're basically these big interpolative databases, interpolative memories. Of course, if you scale up the size of your database and cram more knowledge and patterns into it, you are going to be increasing its performance as measured by a memorization benchmark.</p><p>That's kind of obvious. But as you're doing it, you are not increasing the intelligence of the system one bit. You are increasing the skill of the system. You are increasing its usefulness, its scope of applicability, but not its intelligence because skill is not intelligence. That's the fundamental confusion that people run into. They're confusing skill and intelligence.</p></blockquote><p>Since this interview, models have made enormous progress on Chollet&#8217;s test. OpenAI&#8217;s o3 model, using a high-compute setting and finetuned on a training set, was able to <a href="https://arcprize.org/blog/oai-o3-pub-breakthrough">score 87.5%</a> while at the time of recording, the highest score was only 35%. <em>What explains </em>this improvement is unclear, however. It could have been using reasoning in the Chain of Thought, or better pre-training algorithms, though other people have suggested older models struggled to see the problems.</p><p>The open and important question is what degree of generalisation can we get in the circuitry. As the models get better, the pretraining algorithms get more sample efficient, and the Chain Of Thought&#8217;s get longer, we should probably imagine that generalisation to new problems get better. But how much?</p><h3><strong>What can we take away from the &#8216;stochastic parrot&#8217; saga?</strong></h3><p>Despite what I&#8217;ve just said, I don&#8217;t think most of the &#8220;stochastic parrot&#8221; debate was ever <em>really </em>about circuitry.</p><p>The paper which coined the term was a work of social science, not an investigation into the model&#8217;s internal dynamics. Others carried the term forwards. &#8220;Stochastic parrots&#8221; became part of a broader set of arguments that were part of our desire to explain away the prospect of big change in the world. &#8220;Scaling is over&#8221;; &#8220;the reversal curse means AI is doomed to fail&#8221;; &#8220;they will hit the data wall&#8221;; &#8220;the energy-intensive approach isn&#8217;t the <em>true way</em>&#8221;; &#8220;reasoning will only work in code and math&#8221;, and so on.</p><p>I think it&#8217;s more<em> </em>than just avoiding change though: if it turned out that human intellect was the same as next-token prediction over the Internet, isn&#8217;t that a bit&#8230;dissapointing? Quite a lot of our story for what makes people special depends on Enlightenment ideas about our capacity for reason, our ability to make discoveries, and to use this for progress. If an AI system could do all this too, we'd be set adrift. This is especially so if the ideas are simple: people are sacred, so it follows that their intelligence be mystical and their computation sophisticated? Are we undermined if it is all just simple interpolation over short distances?</p><p>Perhaps it is our fault for attaching ourselves to a set of ideas we understand so poorly. What does it mean to <em>reason? </em>What does it mean to<em> understand? </em>What does it mean to <em>be original</em>? I don&#8217;t really know. As this essay puts it, perhaps <a href="https://www.felixstocker.com/blog/gwh">&#8220;everything is the bar scene in Good Will Hunting&#8221;</a> and we&#8217;re all stochastic parrots reciting obscure passages and contending things like a first year grad student. The essay concludes&#8230;</p><blockquote><p>I guess my best answer to all this is to try to achieve a sort of meta-recognition of your own unoriginality, while still persisting in it. If you are a first-year grad student, and you find yourself making the contention of a first-year grad student, for fuck&#8217;s sake just <em>stop</em>, not least because language models can probably do it better and faster. But if you&#8217;ve taken into account your bounded experience, the determined nature of your reading and the limits to your self-expression, and you still think it&#8217;s worth putting on paper, then by all means, go ahead!<br><br>I think it&#8217;s a bit like conversation at parties; in my first year after university, we all talked about the same stuff - &#8220;are you enjoying your investment banking job? Oh, you went to bed at 3am last night? You&#8217;re also thinking of going to play for your old college rugby team next weekend?&#8221; Now, everyone&#8217;s like &#8220;Did you see they got engaged? I can&#8217;t believe it, she&#8217;s still so young; I&#8217;m so over Hinge dates, I just want my friend to introduce me to someone&#8221;; soon it&#8217;ll be, like, &#8220;I&#8217;m thinking of buying a house; maybe we&#8217;ve had enough of London, we just need more space&#8221;; I can just imagine the agonising over whether you should send your kids to private school. All of this is deeply unoriginal - determined entirely by our job, age, social status, location - and yet is it so bad? Maybe we should all just talk and write a bit more, and never mind what Will Hunting would say about it.</p></blockquote><p>Whatever the answer is, we should probably start looking. (Or at least, I should&#8212;I&#8217;ve just told you about someone else&#8217;s research and someone else&#8217;s essay.) When powerful AI gets made, it&#8217;ll be an unwelcome look at our own specialness and we&#8217;ll need new and better ideas for what this is. These questions are still avoidable&#8212;AI isn&#8217;t changing that much yet&#8212;but at some point, we&#8217;ll have wished we started looking sooner.</p><p>The parrot is dead. Don&#8217;t be the shopkeeper.</p><div><hr></div><p><em>Thanks to Theo Horsley for invaluable comments on drafts of this piece.</em></p>]]></content:encoded></item><item><title><![CDATA[Will there be extreme inequality from AI?]]></title><description><![CDATA[.]]></description><link>https://inferencemagazine.substack.com/p/will-there-be-extreme-inequality</link><guid isPermaLink="false">https://inferencemagazine.substack.com/p/will-there-be-extreme-inequality</guid><dc:creator><![CDATA[Jack Wiseman]]></dc:creator><pubDate>Sat, 05 Apr 2025 19:17:24 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c61df20-b545-4c7b-9acb-75940496383f_1010x1010.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>There are two scenarios which some people fear could cause extreme inequality from AI.</p><p>The first is that automation causes some people&#8217;s wages to diverge dramatically from others. Some jobs involve leveraging AI, so the top performers can get paid a lot more than people doing work which doesn&#8217;t require AI.</p><p>The second is that, at some point, AI and robotics become capable of doing all tasks better than humans can. Humans would have to compete against faster, cheaper, <em>&#8216;better&#8217; </em>machines, and would be unable to keep up. In this scenario, all of the income flows to the owners of capital and the humans who don&#8217;t own capital&#8230;err&#8230;wouldn&#8217;t do very well.</p><p>Economists would talk about these scenarios as &#8216;the distribution within labour&#8217;s share of income&#8217; and &#8216;capital versus labour&#8217;s share of income&#8217;.</p><p>There is precedent for the first scenario&#8212;technologies have changed the structure of the labour market many times over&#8212;but the second scenario requires more first principles reasoning. I am broadly optimistic that we can achieve a good outcome in both scenarios.</p><p><strong>How have previous technological revolutions affected labour markets?</strong></p><p>We can categorise technologies by whether they are mainly <em>a substitute for skilled labour</em>, or <em>a complement</em>. To generalise, technologies of the 19th century were substituting for skilled labour. Power looms and spinning machines replaced artisanal weavers with lower-skilled machine operators. Machine tools and manufacturing displaced craftsmen in lots of goods production too. On the other hand, technologies of the 20th century were generally <em>complements </em>to skilled labour: jobs that were downstream of electrification typically required a high school-level education, while the jobs downstream of computerisation typically required college education.</p><p>The economists Claudia Goldin and Lawrence Katz have established a framework to explain income inequality in terms of the relative pace of technological development and educational attainment. (Their excellent book is aptly-named <em><a href="https://www.amazon.co.uk/between-Education-Technology-Claudia-Goldin/dp/0674035305">The Race Between Education and Technology</a></em>.) To summarise, <em>skill-biased technological change </em>&#8212; like electricity and computers &#8212; is creating new demand for skill. In periods where technological development outpaces improvements in human capital, the pool of workers with suitable skills is growing slower than the demand for their skills. This means their wages rise relative to those without. Conversely, when the supply of skilled labour outpaces new demand for skill, the wage premium shrinks.</p><p>Goldin and Katz map this onto inequality through the 20th century. Inequality decreases for the first three quarters and rises in the final quarter, roughly to the level it began the period. This is congruent with periods of educational acceleration&#8212;the growth of the high school movement in the first third of the century, and the growth of state colleges following the GI bill&#8212;and periods of educational stagnation, from about 1970 onwards.</p><p>This educational expansion also explains why the 20th Century was the American Century. From much earlier in the century, the US was educating <em>a greater portion of its citizens for longer</em> than its European counterparts. In 1960, just 15% of British 17 year-olds were in full time education, while 69.5% of Americans in the same age group were graduating high school. US education was egalitarian, British education was elitist.</p><p>So technology, acting alone, doesn&#8217;t create labour market inequality. Technology is just the demand side of the equation. Education is the supply side.</p><p><strong>How does this relate to AI?</strong></p><p>AI creates an enormous demand for skill.</p><p>First, in <em>using</em> the models. There is huge variety in the quality of a model&#8217;s output depending on the usefulness of its prompt. Some people have strong intuition for where the models excel, how they can be pushed, and where they struggle.</p><p>At the moment, the models are limited by the horizon length they can act for. Deep Research can write a report in five or ten minutes that would take a human about four hours to assemble. But a <a href="https://arxiv.org/abs/2503.14499">recent paper</a> from METR, a model evaluator, has shown that on a large suite of software engineering tasks the time horizon models can act for is doubling every seven months. Were this trend to continue, 2028&#8217;s agents would be able to act for a &#8216;week-equivalent&#8217; of human work. 2030&#8217;s agents would be able to act for a &#8216;month-equivalent&#8217;. (Whether this can generalise outside of software engineering and when this might slow down is uncertain.) </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!eK5N!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F190715f9-88d0-4a57-92cb-ff6f7d4b5f22_720x430.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!eK5N!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F190715f9-88d0-4a57-92cb-ff6f7d4b5f22_720x430.png 424w, https://substackcdn.com/image/fetch/$s_!eK5N!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F190715f9-88d0-4a57-92cb-ff6f7d4b5f22_720x430.png 848w, https://substackcdn.com/image/fetch/$s_!eK5N!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F190715f9-88d0-4a57-92cb-ff6f7d4b5f22_720x430.png 1272w, https://substackcdn.com/image/fetch/$s_!eK5N!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F190715f9-88d0-4a57-92cb-ff6f7d4b5f22_720x430.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!eK5N!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F190715f9-88d0-4a57-92cb-ff6f7d4b5f22_720x430.png" width="720" height="430" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/190715f9-88d0-4a57-92cb-ff6f7d4b5f22_720x430.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:430,&quot;width&quot;:720,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!eK5N!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F190715f9-88d0-4a57-92cb-ff6f7d4b5f22_720x430.png 424w, https://substackcdn.com/image/fetch/$s_!eK5N!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F190715f9-88d0-4a57-92cb-ff6f7d4b5f22_720x430.png 848w, https://substackcdn.com/image/fetch/$s_!eK5N!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F190715f9-88d0-4a57-92cb-ff6f7d4b5f22_720x430.png 1272w, https://substackcdn.com/image/fetch/$s_!eK5N!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F190715f9-88d0-4a57-92cb-ff6f7d4b5f22_720x430.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://metr.org/blog/2025-03-19-measuring-ai-ability-to-complete-long-tasks/">Source</a>, and my earlier post on this <a href="https://jackwiseman.substack.com/p/two-frameworks-for-thinking-about">here</a></figcaption></figure></div><p>But the general trend would provide enormous leverage to knowledge workers who know how to use the models best. A good intuition for this is the Archimedes line, &#8220;give me a lever long enough, and I will move the world&#8221;. Well, the length of the stick is doubling every seven months. As the manager of a team of agents, knowledge workers will decide what tasks to assign, provide context where the model lacks it, correct the model&#8217;s weaknesses and make taste-based decisions. This might feel like a &#8220;promotion for everyone&#8221;.</p><p>The second source of demand for skill will be in automating particular workflows. To automate a task, we have to build scaffolding for the agent to operate within. Part of this is <em>technology-driven</em>&#8212;the agents are too unreliable for unbounded environments&#8212;and part of this is <em>business-need</em>&#8212;companies want to get their agents to act in deterministic ways, with instructions about when to escalate to a human and so on. But a lot of this depends on having good quality data structure across the whole company. One of the reasons we might have seen fewer customer service agents than we might expect, given model capabilities, is that agents need to have suitable infrastructure to find the answers to the customer&#8217;s query. The plan is something like:</p><ol><li><p>Complete the very difficult organisational change to manage the company&#8217;s information in a way that is legible to AI systems.</p></li><li><p><a href="https://www.anthropic.com/engineering/building-effective-agents">Add AI agent</a>.</p></li></ol><p>Step #1 is much harder than step #2! There will be, in the near term, incredible demand for people who have the skill to do #1 and have the know-how for #2.</p><p>Over time, the agents will need less of this kind of scaffolding. They will become <strong>more sample efficient</strong>, meaning they need to see fewer examples of a task before they can do it. They will have <strong>better memory, </strong>which limits their performance today. They will become <strong>more reliable</strong>, needing fewer guardrails. While they are not, humans will fill in the gaps.</p><p><strong>How do we supply the skill to AI-driven demand?</strong></p><p>One of the criticisms made of the Katz and Goldin book is that it can treat additional years of schooling and skill too monolithically. Whether additional years of education are <em>actually improving </em>skill to the degree we might hope is unclear. <a href="https://en.wikipedia.org/wiki/The_Case_Against_Education">Work</a> from the economist Bryan Caplan has shown that two thirds of the college wage premium is attributable to the signalling value of a degree, and just a third was attributable to human capital improvement. We can&#8217;t just spam the &#8220;more education&#8221; button and hope for better outcomes. At least in the UK, <a href="https://www.bbc.co.uk/news/education-49841620">50% of people are going to university already</a>.</p><p>However, this is the first general-purpose technology that can help us improve directly<strong>. </strong>Electricity only very weakly helps to acquire skills for industrial production&#8212;perhaps by allowing you to read later into the night&#8212;but AI can be a tutor. The quality of education can be radically improved.</p><p>When it comes to inequality, an underrated concern for future wage differences would be that independent schools adopt AI tutoring much faster than state-funded schools. A recent news story highlighted a Texas private school which had been able to boost their test scores to the top 2% in the US. Someone I know who started an AI tutoring company is only selling to microschools in the US, because it would have been slower to sell to public school districts. OpenAI has created <a href="https://openai.com/index/openai-and-the-csu-system/">ChatGPTedu</a> and partnered with individual universities and the California State system to provide free access to students, and Anthropic is <a href="https://www.theverge.com/ai-artificial-intelligence/641193/openai-anthropic-education-tool-college">doing something similar</a>.</p><p>The UK&#8217;s <a href="https://www.gov.uk/government/publications/generative-artificial-intelligence-in-education/generative-artificial-intelligence-ai-in-education">national strategy</a> seems to be to keep AI &#8216;teacher-facing&#8217; in state schools.</p><p><strong>Which jobs are most exposed to AI?</strong></p><p>The <a href="https://arxiv.org/abs/2503.04761">Anthropic Economic Index</a> has some precursory data on exposure by job. They take anonymised Claude interactions and use the models to categorise the content of these conversations. They found people are overwhelmingly using Claude for software engineering tasks.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pnnp!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4d73b7a-ffb9-4323-a1b1-c7b27966bb58_962x878.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pnnp!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4d73b7a-ffb9-4323-a1b1-c7b27966bb58_962x878.png 424w, https://substackcdn.com/image/fetch/$s_!pnnp!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4d73b7a-ffb9-4323-a1b1-c7b27966bb58_962x878.png 848w, https://substackcdn.com/image/fetch/$s_!pnnp!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4d73b7a-ffb9-4323-a1b1-c7b27966bb58_962x878.png 1272w, https://substackcdn.com/image/fetch/$s_!pnnp!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4d73b7a-ffb9-4323-a1b1-c7b27966bb58_962x878.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pnnp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4d73b7a-ffb9-4323-a1b1-c7b27966bb58_962x878.png" width="962" height="878" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c4d73b7a-ffb9-4323-a1b1-c7b27966bb58_962x878.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:878,&quot;width&quot;:962,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!pnnp!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4d73b7a-ffb9-4323-a1b1-c7b27966bb58_962x878.png 424w, https://substackcdn.com/image/fetch/$s_!pnnp!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4d73b7a-ffb9-4323-a1b1-c7b27966bb58_962x878.png 848w, https://substackcdn.com/image/fetch/$s_!pnnp!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4d73b7a-ffb9-4323-a1b1-c7b27966bb58_962x878.png 1272w, https://substackcdn.com/image/fetch/$s_!pnnp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4d73b7a-ffb9-4323-a1b1-c7b27966bb58_962x878.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://www.anthropic.com/news/the-anthropic-economic-index">Source</a></figcaption></figure></div><p>This maps onto Michael Webb&#8217;s <a href="https://www.michaelwebb.co/webb_ai.pdf">prospective forecast</a> which uses semantic analysis of patents to evaluate which jobs are most exposed to AI. He found that the 88th percentile of the wage distribution was most exposed to AI, similar to Anthropic&#8217;s retrospective analysis.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!QmHp!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c6c3229-aa3a-441d-a95a-c232b080921c_1600x1248.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!QmHp!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c6c3229-aa3a-441d-a95a-c232b080921c_1600x1248.png 424w, https://substackcdn.com/image/fetch/$s_!QmHp!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c6c3229-aa3a-441d-a95a-c232b080921c_1600x1248.png 848w, https://substackcdn.com/image/fetch/$s_!QmHp!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c6c3229-aa3a-441d-a95a-c232b080921c_1600x1248.png 1272w, https://substackcdn.com/image/fetch/$s_!QmHp!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c6c3229-aa3a-441d-a95a-c232b080921c_1600x1248.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!QmHp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c6c3229-aa3a-441d-a95a-c232b080921c_1600x1248.png" width="1456" height="1136" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0c6c3229-aa3a-441d-a95a-c232b080921c_1600x1248.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1136,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!QmHp!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c6c3229-aa3a-441d-a95a-c232b080921c_1600x1248.png 424w, https://substackcdn.com/image/fetch/$s_!QmHp!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c6c3229-aa3a-441d-a95a-c232b080921c_1600x1248.png 848w, https://substackcdn.com/image/fetch/$s_!QmHp!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c6c3229-aa3a-441d-a95a-c232b080921c_1600x1248.png 1272w, https://substackcdn.com/image/fetch/$s_!QmHp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c6c3229-aa3a-441d-a95a-c232b080921c_1600x1248.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://www.michaelwebb.co/webb_ai.pdf">Source</a></figcaption></figure></div><p>The Anthropic analysis also found that 57% of queries were augmentative (&#8220;help me do this thing&#8221;) while 43% of queries were automation (&#8220;do this thing&#8221;).</p><p>The relevance of this for inequality is how exposure to AI <em>changes the returns to talent</em>. There are some domains where AI augmentation can &#8216;raise the performance floor&#8217; by mitigating the weaknesses of the lowest skill employees but cannot meaningfully uplift the highest skill performers. <a href="https://www.nber.org/papers/w31161">This paper</a> finds this to be true for call centre workers, supported by an LLM-based system&#8212;you can only be <em>so good </em>at answering the customers&#8217; query. In some cases, using AI systems has actually decreased the performance of some high skill workers. An analysis of <a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5162111">legal work found this</a>. However, in some domains, and perhaps software engineering the returns to greater talent are &#8216;uncapped&#8217;. The absolute top performers can be many times more effective than others.</p><p>When the constraint becomes <em>how many agents </em>can each person manage to do more, or <em>teams of agents</em>, the returns to talent can be magnified in some areas. The less scarce these skills can be, the lower the wage divergence can be.</p><p><strong>What follows </strong><em><strong>cognitive </strong></em><strong>automation?</strong></p><p>Until we make progress on robotics, AI automation will be limited to cognitive labour. This can still cause the physical world to change enormously: AI systems can better organise production and can &#8216;deskill&#8217; tasks that humans would otherwise do. Science is organised around principal investigators in the same way that steam-powered factories were organised around a drive belt.</p><p>Even at near-complete cognitive automation, output would remain bottlenecked by human capacity for real world tasks. Baumol&#8217;s cost disease means wages for physical tasks will be much higher than today.</p><p>To some extent, cognitive automation can help us make faster progress in robotics. A digital robotics researcher could design better experiments, create simulated environments to gather training data, and develop more efficient algorithms for training and inference. This would accelerate robotics progress, but we are still far away from robots that could do <em>all </em>the tasks that humans do.</p><p><strong>We wouldn&#8217;t automate all tasks in the economy</strong></p><p>In the second scenario, where AI and robots are <em>capable </em>of doing all tasks in the economy, I am unconvinced that humans would be left with nothing to do. This is for a few reasons:</p><ul><li><p>First, there will be a long period where some types of context which humans have that models lack. Hayek&#8217;s essay <em><a href="https://www.econlib.org/library/Essays/hykKnw.html">The Use of Knowledge In Society</a> </em>makes the point that there are local and temporary forms of knowledge which cannot be captured by any central system.</p></li><li><p>Second, humans will retain a preference for interacting with other humans. <em>Especially because</em> human labour gets so expensive, goods and services made with human labour become positional goods. People will still do work which require human-to-human trust and connection.</p></li><li><p>Third, we aren&#8217;t going to give AI systems legal personhood. There is no &#8220;justice&#8221; system for AI agents and so there cannot be consequences for actions AI systems take in the world. <em>Someone </em>is going to have to be ultimately responsible.<br><br>Part of this is that humans want the division of responsibility. Sometimes an executive&#8217;s job is to do things that help the company, but another component of their job is so that if something goes wrong in their domain, the CEO can turn to the board and say, &#8220;Well, I hired this person who is credible and was supposed to be responsible, so it&#8217;s not my fault.&#8221;</p></li><li><p>Fourth, people will create new jobs in bureaucracies. Yale University has <a href="https://yaledailynews.com/blog/2021/11/10/reluctance-on-the-part-of-its-leadership-to-lead-yales-administration-increases-by-nearly-50-percent/">nearly a 1:1 ratio of administrators to undergraduates</a>. Especially as people get richer, they tend to value safety more, so there will be no limit to the amount of things we can make up for humans to do.</p></li><li><p>And finally, people can lobby governments to step in to create jobs or make it basically impossible for companies to fire people.</p></li></ul><p>Based on these factors, some jobs will continue to be done by humans, and so it will be possible to retain a balance within the labour share of income. To the extent that human labour remains a complement to capital, because of the factors mentioned above, then labour will retain an equivalent share of income. The idea that &#8216;capital&#8217; will dominate labour&#8217;s share of income (i.e. capital will take all of the gains) depends on the idea that AI systems and robots will be perfect substitutes for humans. Nowhere do humans retain a comparative advantage.</p><p>One way to think about this is to imagine, in 1800, if you saw all the mechanisation coming, surely you would assume that &#8216;capital&#8217; becomes an enormous fraction of the economy, but it didn&#8217;t and things remained in equilibrium because wages rose too. Everything balances out, just at much greater equilibriums, so long as labour remains a complement to capital (or the rules arbitrarily enforce this should be the case).</p><p>I expect this second scenario will take a long time to come to pass, much longer than most people in AI, for reasons discussed <a href="https://inferencemagazine.substack.com/i/155018281/cognitive-labour-will-be-automated-before-physical-labour-and-could-be-automated-much-more-quickly-than-previous-technological-revolutions">here</a>. Overall, I think the picture is optimistic. It ultimately hinges on your view of human nature &#8212; how much do we value the <em>humanity </em>of other people in our transactions? When people get richer, they buy fairtrade and other &#8220;ethical&#8221; products. People do care about the provenance of positional goods, and they do care to watch other humans race cars around tracks, based on arcane rules, to watch sports and chess. I expect, and hope, this remains in the future.</p><p>If we can accelerate educational attainment to give as many people the skills for AI as possible, we should avoid a future with greater inequality.</p>]]></content:encoded></item><item><title><![CDATA[Coreweave]]></title><description><![CDATA[A quirk of human desire]]></description><link>https://inferencemagazine.substack.com/p/coreweave</link><guid isPermaLink="false">https://inferencemagazine.substack.com/p/coreweave</guid><dc:creator><![CDATA[Jack Wiseman]]></dc:creator><pubDate>Tue, 01 Apr 2025 23:23:23 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c61df20-b545-4c7b-9acb-75940496383f_1010x1010.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>CoreWeave is the largest AI neocloud and went public last week. Some commentators have wanted to use the CoreWeave IPO as a way to cast doubt on the &#8216;Generative AI&#8217; industry. See, for example, a recent Bloomberg Opinion column entitled <em><a href="https://www.bloomberg.com/opinion/articles/2025-03-28/coreweave-s-ipo-will-expose-ai-s-dirty-secrets">CoreWeave&#8217;s IPO Will Expose AI&#8217;s Dirty Secrets</a></em>&#8230;</p><blockquote><p>CoreWeave stands to be a bellwether for the AI industry as a whole &#8212; a must-watch stock as questions about return on investment grow ever louder. Any slowdown in demand for CoreWeave&#8217;s &#8220;compute,&#8221; as the term goes, will be seen by Wall Street as a heavy indication of a softening across the board, dragging down Amazon.com Inc., Google parent Alphabet Inc., Microsoft Corp., Nvidia and several others.</p></blockquote><p>This isn&#8217;t true at all. CoreWeave is a weird company: it got 62% of its revenue last year from Microsoft, and another 15% from its biggest supplier, Nvidia. This makes it an interesting lens into compute markets, but it cannot be an <em>index of them</em>. The best intuition I have for the company was from a line in Scott Alexander&#8217;s <em>Meditations on Moloch&#8230;</em></p><blockquote><p>Las Vegas doesn&#8217;t exist because of some decision to hedonically optimize civilization, it exists because of a quirk in <a href="http://journal.frontiersin.org/Journal/10.3389/fnbeh.2013.00206/full">dopaminergic reward circuits</a>, plus the microstructure of an uneven regulatory environment, plus Schelling points.</p></blockquote><p>Just as the casinos and hotels we&#8217;ve carved out in the desert are the result of a quirk in human desire, Coreweave is carved out too: from Nvidia&#8217;s desire to reduce its customer concentration, and Microsoft&#8217;s desire to grow its asset base efficiently. Las Vegas was on track to become an irrelevant, middle-of-nowhere railroad town; but historical contingency intervened. In 2022, CoreWeave appeared to be an unprofitable Ethereum miner, but circumstances&#8212;AI progress&#8212;and a whole lot of capital, intervened.</p><p>One way of looking at the company is as a business line for Nvidia. Nvidia owns 6% of the equity, but this understates the depth of the relationship: CoreWeave is arranging tens of billions of dollars in credit facilities, to spend the majority on Nvidia chips and networking, before renting some fraction of this compute back to Nvidia. This could be, partially, a convenience for their internal R&amp;D efforts&#8212;it's a hassle to build their own datacentres&#8212;but this doesn&#8217;t feel like a sufficient explanation, because Nvidia already runs its own small cloud provider.</p><p>It is, in part, an effort to weaken the bargaining power of their large customers. About half of Nvidia&#8217;s revenue comes from just four customers&#8212;Amazon, Google, Meta, and Microsoft&#8212;all of whom have internal chip design efforts, so they are vulnerable to these companies changing their orders. This would explain why CoreWeave was the first to offer Nvidia&#8217;s newest hardware, the GB200, in February 2025. But this explanation also feels insufficient: if Microsoft is a majority of the revenue, it's hardly diminishing Microsoft exposure <em>that much</em>.</p><p>Nvidia benefits from CoreWeave existing. When demand for CoreWeave&#8217;s IPO looked shaky, Nvidia backstopped it with an additional $250 million investment. The CoreWeave CEO said <a href="https://www.bloomberg.com/news/articles/2025-03-28/coreweave-s-debut-dud-extends-ipo-malaise-instead-of-ending-it">they couldn&#8217;t have done it without them</a>.</p><p>The other lens is through Microsoft. For them, CoreWeave is a tool to manage their datacentre fleet construction. When a cloud provider wants to build a new datacentre, they are looking at about five years to get it running. If you need to build new power, the decision timeline is even further out. Microsoft&#8217;s decision to restart a nuclear reactor at Three Mile Island, will start providing power in 2028, and they have a 20-year power purchase agreement. How could they decide that this would be a good investment last year? It is extremely difficult to predict demand on this horizon; you&#8217;d need to answer questions like:</p><ul><li><p>How much will hardware improve, in energy efficiency terms?</p></li><li><p>How much will software improve, in inference cost per token terms?</p></li><li><p>How many tokens will we want to spend for each query, on average?</p></li><li><p>How many queries will we want to make, if we have long-horizon agents, or an open-ended AGI?</p></li><li><p><em>Where</em> will we want to do inference in the world, so datacentres can be nearby for the lowest latency? (As Satya <a href="https://www.dwarkesh.com/p/satya-nadella">put it</a>, &#8220;[a]t the end of the day, speed of light is speed of light, so you can't have one data center in Texas and say, &#8216;I'm going to serve the world from there.&#8217;&#8221; Clearly a subtle jibe at OpenAI&#8217;s expectation that Stargate can service <a href="https://www.theinformation.com/articles/openai-forecast-shows-shift-from-microsoft-to-softbank?rc=u28gfh">three quarters of their compute needs</a>.)</p></li></ul><p>The fortunate thing is that, if you get this prediction wrong, some of the investment can be repurposed. The land, power, and datacentre that was meant to be for AI training can be repurposed for AI inference, Azure CPUs, or storage. Microsoft said five times in their recent earnings that they are building a &#8220;fungible fleet&#8221;, and they also mentioned that about half of their $80 billion in AI capital expenditure is being spent on long-lived assets, and the other half is short-lived assets, like GPUs.</p><p>This is where CoreWeave comes in. It would be ideal for Microsoft if Azure owned all the compute they need for AI inference to serve all their customers in 2029, because the margins on this are better. But doing this requires taking risk on long-lease assets like power-purchase agreements and datacentres. Signing a long-term contract with CoreWeave means they can have access without needing to take on the risk of the long-lease assets. (Renting a piece of hardware for almost its entire useful life is as good as owning it.) You can think of Microsoft&#8217;s own fleet as &#8216;baseload&#8217; compute, which they are more confident they can make a return on, and Coreweave as &#8216;top up&#8217;, for which they accept a reduced margin to ensure they can serve demand without long-term risk. The same pattern will be true for OpenAI.</p><p>The challenge then, is finding a price where this works for both sides. The limited useful life of hardware is a strain on this. The cycle which is going to dominate rentals will be:</p><ol><li><p>A new generation of hardware is released, and renters sign up to 2-3 year contracts, while others sign shorter-term 1 year or 6 month deals.</p></li><li><p>As these contracts expire, the next generation of hardware is around the corner, with better cost efficiency and energy efficiency.</p></li><li><p>This exerts two downward forces on the rental price for this hardware. First, there&#8217;s simply a glut for short leases, because none of that hardware is being signed to longer leases. And second, when the newer hardware is more performant, the incentive to switch is stronger.</p></li><li><p>Over time, the marginal cost of <em>operation </em>for old hardware will be higher than both the <em>upfront and operating </em>cost<em> </em>of new hardware (in compute per dollar). Jensen also highlighted on Nvidia&#8217;s latest earnings call that there&#8217;s an opportunity cost for datacentre space and power too:</p></li></ol><blockquote><p>If you have a 100-megawatt data center, if the performance or the throughput in that 100-megawatt or the gigawatt data center is four times or eight times higher, your revenues for that gigawatt data center is eight times higher. And the reason that is so different than data centers of the past is because AI factories are directly monetizable through its tokens generated.</p></blockquote><p>SemiAnalysis has <a href="https://semianalysis.com/2024/04/10/nvidia-blackwell-perf-tco-analysis/#gpt-4-profitability-cost-inference-simulator-parallelism-explained-performance-tco-modeling-in-large-small-model-inference-and-training">specific numbers on this improvement in &#8216;performance against ownership cost&#8217;</a>, though it's behind their paywall so I won&#8217;t quote them here. Satya gave a rule of thumb on the last Microsoft earnings call about how this dynamic affects their investment decision&#8230;</p><blockquote><p>You don&#8217;t want to buy too much of anything at one time because of the Moore&#8217;s Law every year is going to give you 2x. Your optimization is going to give you 10x. You want to continuously upgrade the fleet, modernize the fleet, age the fleet, and at the end of the day, have the right ratio of modernization and demand-driven monetization to what you think of as the training expense.</p></blockquote><p>At Nvidia GTC, Jensen suggested <a href="https://semianalysis.com/2025/03/19/nvidia-gtc-2025-built-for-reasoning-vera-rubin-kyber-cpo-dynamo-inference-jensen-math-feynman/">even more aggressive rates of improvement</a> and went as far to claim, &#8220;When Blackwells start shipping in volume, you couldn&#8217;t even give Hoppers away.&#8221; This annual release cycle compresses the hardware&#8217;s effective life, which is an issue for CoreWeave. Their IPO filings say they breakeven on hardware on average 2.5 years after purchase, but this will have been buoyed by the H100 shortage in the earliest days of the ChatGPT-induced investment boom and the delays to the new Blackwell generation. SemiAnalysis <a href="https://semianalysis.com/2025/03/26/the-gpu-cloud-clustermax-rating-system-how-to-rent-gpus/#applying-the-semianalysis-tco-framework-to-coreweave">has a model for CoreWeave&#8217;s payback period</a> again, which I won&#8217;t quote specifically, but it seems very difficult to imagine if Nvidia can keep its pace of releases and improvements as high as it intends to, that CoreWeave would be able to earn a suitable return in time.</p><p>This is what makes <a href="https://www.reuters.com/technology/artificial-intelligence/coreweave-strikes-12-billion-contract-with-openai-ahead-ipo-sources-say-2025-03-10/">OpenAI</a>&#8217;s 5-year contracts with CoreWeave so interesting. Given the rate of improvement in hardware, it seems undesirable to commit to use 2025 accelerators in 2030. What might be going on here?</p><ul><li><p>This might not be for specific hardware, but for capacity, in FLOP terms, or otherwise.</p></li><li><p>They might be able to get out of these contracts. (It isn&#8217;t so clear how long Microsoft&#8217;s contracts were, but there is <a href="https://www.ft.com/content/f3d9d339-42ef-4979-bf52-89ecd699dea2">an FT report</a> they&#8217;ve been able to step back from some capacity. Note that CoreWeave disputes this.)</p></li><li><p>Perhaps they are willing to pay for 2025 hardware in 2030, because they know CoreWeave has earliest access to the newest hardware which offers a few months of lower margins.</p></li><li><p>Or something else&#8230;</p></li></ul><p>While we don&#8217;t know for sure, the overall picture is clear: CoreWeave&#8217;s existence doesn&#8217;t depend on an underlying economic engine, but whether it is advantageous to Nvidia and the clouds (including OpenAI). There are a lot of other peculiarities that I&#8217;ve left to the side to make this point: CoreWeave&#8217;s subsidiaries borrow money to build compute with loans that are <em>secured on the compute </em>which will be worth very little when the debt comes due. CoreWeave had a <a href="https://www.ft.com/content/cb94eb68-ccb5-4fb3-b903-0aae17b836dd">technical default on a loan because of an admin error</a>. CoreWeave&#8217;s founders, per <a href="https://www.thediff.co/archive/the-coreweave-triangle/#fn2">The Diff</a>, have sold $450 million in secondaries, and now own just 2.4% of the equity. Finally, <a href="https://www.wheresyoured.at/core-incompetency/">this source</a> casts doubt on CoreWeave&#8217;s ability to grow its power supply through a partner, Core Scientific.</p><p>At the end of the day, none of these issues answer CoreWeave&#8217;s main question: does its existence provide convenience to Nvidia, Microsoft, and OpenAI? If it goes bust in the next few years, this won&#8217;t reflect the top of an AI bubble, moreso that it stopped making sense to prop it up.</p><h2>Otherwise</h2><p><strong>OpenAI raised $40 billion at $300 billion post-money valuation</strong>. Lots of people will be shocked by the valuation &#8212; Anthropic, by comparison, raised at $60 billion &#8212; but it is further confirmation that <em>research labs are becoming product companies</em>. When Sam Altman was <a href="https://stratechery.com/2025/an-interview-with-openai-ceo-sam-altman-about-building-a-consumer-tech-company/">interviewed on Stratechery</a>, Ben asked, &#8220;What&#8217;s going to be more valuable in five years? A 1-billion daily active user destination site that doesn&#8217;t have to do customer acquisition, or the state-of-the-art model?&#8221;; Sam&#8217;s response, &#8220;The 1-billion user site I think.&#8221; You can simplify this investment in OpenAI to a simple question: can this company become Meta? In this light, Meta is about five times bigger than OpenAI today, in market cap terms, so a 20% chance seems about right.</p><p><strong>xAI bought X (formerly Twitter) in a transaction valuing xAI at $80 billion and X at $33 billion</strong>. The story people would like to tell here is that this transaction makes sense most of all for Elon: to offload X&#8217;s debt onto xAI, which has a lower cost of capital. Matt Levine <a href="https://www.bloomberg.com/opinion/newsletters/2025-03-31/musk-merged-his-xes?srnd=undefined">will cover this transaction</a> better than I can, but it is worth noting that, in light of Sam&#8217;s comments above, that 600 million weekly active users is more than OpenAI&#8217;s 500 million (though OpenAI is growing <em>much </em>faster).</p><p><strong>Anthropic released two papers on the thought patterns of language models. </strong><em><a href="https://transformer-circuits.pub/2025/attribution-graphs/methods.html">Circuit Tracing: Revealing Computational Graphs in Language Models</a></em> and <em><a href="https://transformer-circuits.pub/2025/attribution-graphs/biology.html">On the Biology of a Large Language Model</a>.</em></p><p><strong>I wrote about why models of explosive economic growth, like Epoch&#8217;s GATE model, are misleading <a href="https://substack.com/home/post/p-159632770">here</a>.</strong></p>]]></content:encoded></item><item><title><![CDATA[The uphill battle to “mitigate the risks”]]></title><description><![CDATA[The EU's Code of Practice reveals we're all unsure how to regulate AI.]]></description><link>https://inferencemagazine.substack.com/p/the-uphill-battle-to-mitigate-the</link><guid isPermaLink="false">https://inferencemagazine.substack.com/p/the-uphill-battle-to-mitigate-the</guid><dc:creator><![CDATA[Jack Wiseman]]></dc:creator><pubDate>Mon, 17 Mar 2025 00:48:41 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c61df20-b545-4c7b-9acb-75940496383f_1010x1010.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The go-to slogan for AI aficionados&#8212;and the tagline of this very magazine&#8212;is that we should "capture the benefits, and mitigate the risks". Its essential qualities are, one, that it acknowledges pros and cons, but, two, is sufficiently abstract that everyone can assign their own meaning. We all agree!</p><p>At some point, it would be preferable to have more concrete consensus on what we actually believe. The CEOs from the three leading labs <a href="https://www.safe.ai/work/statement-on-ai-risk">have said</a> extinction risk from AI should be treated on par with &#8220;other societal-scale risks such as pandemics and nuclear war&#8221; and all the labs have established frontier safety frameworks. Now comes the EU&#8217;s clarification: the <a href="https://digital-strategy.ec.europa.eu/en/library/third-draft-general-purpose-ai-code-practice-published-written-independent-experts">third draft</a> of their Code of Practice was released last week. This will implement the general-purpose aspects of the EU AI Act, passed this time last year.</p><p>Reading the draft is pleasantly surprising. There are no crazy requirements that caricatures of EU digital regulation might imply. On the surface, each request seems fairly sensible. However, once again, it has deferred the toughest questions. So often, the requirements are set as &#8220;appropriate&#8221; rather than specified. In the Safety and Security section, the word appears 107 times. Who will decide what this means?</p><p>There is a saying that in a democracy, a government must be satisfied that any laws they make can be enforced by their opponent. Perhaps there is a corollary here: if one writes &#8220;appropriate requirements&#8221; in an EU implementation document, they must be satisfied with the definition being set, not by the talented authors of the Code, but by a junior Brussels technocrat. The same kind who specified a training compute threshold of 1e25 FLOP in the original Act.</p><p>While the option value of flexibility might be preferable in the short term, this cedes too much power to the regulator and creates too much uncertainty for labs in the future.</p><div><hr></div><p>This is clear in the section on systemic risk requirements. At a high level, these requirements are aiming to say, &#8220;If we observe this [sign of a bad thing], then we can [pull this handle].&#8221; This might mean, &#8220;If the model shows evidence of deceiving us as to its true intentions, we can pause training and investigate&#8221;, or &#8220;If the model helps a novice do harmful synthetic biology 5 times faster than they would be able to just Googling, we would harden lab security before continuing training, and improve model robustness, before deploying&#8221;. All reasonable requests. The challenge, however, is that <strong>we are leaving the regime where it was cheap&#8212;both in computational resources and time&#8212;to elicit model capabilities.</strong></p><p>To look at this concretely, the Code of Practice requires:</p><blockquote><p>[Signatories shall]</p><p><strong>assess and, as necessary, mitigate systemic risks at appropriate milestones that are defined and documented before training starts</strong>, where systemic risks stemming from the model in training could materially increase, such as:</p><p>training compute based milestones (e.g. every two- to four-fold increase in effective compute);</p><p>development process based milestones (e.g.: during or after phases of fine-tuning or reinforcement-learning; before granting more individuals access to the model; or before granting the model more affordances such as network, internet, or hardware access); or</p><p>metrics based milestones (e.g. at predetermined levels of training loss or evaluation performance)</p><p>implement <strong>appropriate procedures to identify substantial changes in systemic risks which warrant pausing development</strong> to conduct further systemic risk assessment, such as automated benchmarks enabling a highly scalable and real-time identification of capability increases thereby lowering the risk of human or organisational bottlenecks.<br><br>[emphasis mine]</p></blockquote><p>For older models, we can use &#8216;proof by non-example&#8217;: run GPT-3, ask it multiple-choice biology questions, see that it isn&#8217;t good enough to help with synthetic biology compared to just browsing the Internet, and, by induction, it is safe to deploy. This is also very cheap! Getting the model to answer these questions does not cost much, and the computer can handle the marking too.</p><p>This cannot be the case forever. Take <a href="https://x.com/hud_zah/status/1880353827771076947">this example</a>: using Claude 3.5 Sonnet, a college student built a nuclear fusor in his kitchen in 36 hours. While this is not actually <em>that </em>dangerous&#8212;most of the information is Google-able&#8212;it is a toy example that demonstrates the kind of &#8216;human uplift&#8217; we might care to study. &#8220;How much support does a model provide novices doing engineering tasks that might take days unassisted?&#8221;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!SYqJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ab37b7a-f487-43ee-a590-3dc078d86f12_544x582.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!SYqJ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ab37b7a-f487-43ee-a590-3dc078d86f12_544x582.png 424w, https://substackcdn.com/image/fetch/$s_!SYqJ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ab37b7a-f487-43ee-a590-3dc078d86f12_544x582.png 848w, https://substackcdn.com/image/fetch/$s_!SYqJ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ab37b7a-f487-43ee-a590-3dc078d86f12_544x582.png 1272w, https://substackcdn.com/image/fetch/$s_!SYqJ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ab37b7a-f487-43ee-a590-3dc078d86f12_544x582.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!SYqJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ab37b7a-f487-43ee-a590-3dc078d86f12_544x582.png" width="682" height="729.6397058823529" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4ab37b7a-f487-43ee-a590-3dc078d86f12_544x582.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:582,&quot;width&quot;:544,&quot;resizeWidth&quot;:682,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!SYqJ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ab37b7a-f487-43ee-a590-3dc078d86f12_544x582.png 424w, https://substackcdn.com/image/fetch/$s_!SYqJ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ab37b7a-f487-43ee-a590-3dc078d86f12_544x582.png 848w, https://substackcdn.com/image/fetch/$s_!SYqJ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ab37b7a-f487-43ee-a590-3dc078d86f12_544x582.png 1272w, https://substackcdn.com/image/fetch/$s_!SYqJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4ab37b7a-f487-43ee-a590-3dc078d86f12_544x582.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Applying the EU&#8217;s rules for systemic risk to this example, would we have to stop training the model at multiple milestones&#8212;&#8220;appropriate&#8221;, as determined by the regulator&#8212;and see how much faster it helps a novice to make a nuclear fusor? The rules say it is permissible to use automated benchmarks, but there aren&#8217;t any automated tests that could answer this.</p><p>It would only be possible to show a multiple-choice score from a set of questions about nuclear fusors, that boils down to, &#8220;this model knows a lot about nuclear fusors&#8221;. From this score, the researcher&#8217;s reactions would be unremarkable. <em>Of course</em> the model picked up knowledge about this, the models are good at memorising the Internet. And second, a high score on this benchmark is slightly meaningless: I don&#8217;t expect it would cause anyone to actually stop training to take mitigations.</p><p>Therefore, what would cause someone to stop training? Do we have to run red teaming at each milestone, with a new college student each time trying to build a nuclear fusor as quickly as they can? Is there an in-between, where we can create a series of tasks or environments that simulate aspects of aiding a human with this task? (This all assumes that we could <em>foresee </em>that we&#8217;d need to evaluate this capability, but some capabilities can make discontinuous progress. Imagine you&#8217;d be planning coding evaluations for the state of the art 12 months ago, it seems quite likely you&#8217;d have undershot model capabilities. I am speculating but I don&#8217;t think we knew it&#8217;d be possible to make a nuclear fusor in 36 hours when Sonnet 3.5 was released.)</p><p>The second challenge for these interventions is <strong>how much more expensive they are in computation</strong>. As <a href="https://inferencemagazine.substack.com/p/how-much-economic-growth-from-ai">we&#8217;ve written previously</a>, all the labs&#8217; R&amp;D will be compute-constrained, and so any compute dedicated to running evaluations is not running experiments. As multiple-choice questions become less useful, we will need to run more long-horizon tests. For example, &#8220;Can this system write an expert-level plan for building a nuclear fusor and correctly order the components online?&#8221;. For a good evaluation, the researchers would need to run this at quite a large scale, using perhaps 100 copies of the AI system. As before, it is difficult to interpret the results. Let&#8217;s say that just 6 copies are successfully able to order the components and all the others made a mistake, or got stuck in a loop, or something else. Should we pause training <em>then </em>to take an intervention? What about if <em>60</em> copies succeeded? Surely this standard would differ for internal deployment, and deploying to hundreds of millions of people?</p><p>There is a similar problem with evaluating deception during training. Some people believe that the model might conceal its true intentions, reporting that it will be faithful to the values and rules it is taught in training but then rebelling later in training or deployment. (Others don&#8217;t think that this kind of &#8216;loss of control&#8217; risk is possible.) We only have one evaluation for this capability, <a href="https://arxiv.org/pdf/2412.04984">from Apollo</a>. This gives the model a system prompt, and later exposes the developer&#8217;s true goals to the model, which conflict with the system prompt. The test evaluates whether the models continue with their original goal or follow the developer&#8217;s true goals. While valuable, this will not be able to provide the kind of conclusive evidence that would cause a lab to pause their training.</p><p>Later, the Code requires that before models are deployed:</p><blockquote><p>&#8220;model evaluations are performed&#8230;proportionate to the systemic risk assessed to: (1) elicit the upper limit of current and reasonably foreseeable capabilities&#8230; [and] 4) match the realistic model elicitation capabilities of potential misuse actors&#8221;</p><p>&#8220;The given time shall be proportionate to: (a) the magnitude of the systemic risk assessed&#8230;An assessment time of at least 20 business days could, e.g., indicate that model evaluation teams were given enough time for most systemic risks and model evaluation methods.&#8221;</p></blockquote><p>None of this is objectionable, but it is impossible to satisfy these conditions without making organisational overhead go through the roof! Internal teams, under these rules, would need to elicit the full extent of cyber offence capabilities; chemical, biological, radiological, nuclear capabilities; the potential for harmful manipulation; and the potential for loss of control. And for some of these capabilities, it does not mean just interacting with the model as it is, but with extra scaffolding too. That&#8217;s a lot to do in 20 days before deployment! Third party evaluators are given just seven days' access before deployment. It also seems difficult to imagine them having enough time to elicit the full extent of the model&#8217;s capabilities.</p><p>This is a version of the <a href="https://en.wikipedia.org/wiki/Jeep_problem">&#8216;jeep problem&#8217;</a>: the further the jeep goes into the desert, the more fuel it needs to carry, but to deal with its heavier weight it needs to take more fuel. At some point, this becomes prohibitive to going any further! Likewise, as the models get more capable, the range of their dangerous capabilities gets wider (they could do more things) and longer (they are more useful for longer periods), so more and more compute and evaluation time is required, until it causes training to grind to a halt.</p><div><hr></div><p>In some ways, the vagueness of the Code reveals a deep truth: there is not a suitable toolkit to regulate AI development yet. The current proof-by-non-example regime is going to run out of steam, and we don&#8217;t have answers for what will come after. We have to solve for both constraints: training and deployment has to continue with minimal interruption, but we need to elicit the full risks of the models and put in place safeguards. Answering these questions in this versions of the code could lock in an incorrect regime. Also, the EU can leave the door open for lenient enforcement, if they <a href="https://inferencemagazine.substack.com/p/peak-brussels">face pressure from the US</a>, or to leave scope to enforce more stringently later.</p><p>To finish where we started: this doesn&#8217;t seem like a worthwhile trade. The Code of Practice cedes almost complete power to the EU AI Office to decide what is &#8220;appropriate&#8221;. They could be pausing training very often for extremely long tests to confirm it is safe to continue. This kind of error is the same as killed nuclear power: the International Commission for Radiological Protection has principles to be &#8220;precautionary&#8221; and &#8220;prudent&#8221; but this is poorly specified and has cascaded through the regulatory states&#8217; poor incentives in the UK and the US. Now, the UK over-regulates radiation by a factor of 100 and struggles to build new power stations in 25 years. The same cannot happen to AI.</p><p>While the authors of the Code are well-meaning, and genuinely proportionate, the standard we should judge the code to is whether the junior official who will implement these rules in 2, 5, or 25 years time will do so with the same spirit. The constraints on the authors are enormous: satisfying proportionate constraints on training, against an immature scientific discipline for eliciting dangerous capabilities, and balancing the geopolitical headwinds that EU enforcement faces. These challenges, however, cannot justify complete discretion to the AI Office.</p>]]></content:encoded></item><item><title><![CDATA[The Masters of Our Destiny]]></title><description><![CDATA[Technological Sovereignty in the Compressed 21st Century]]></description><link>https://inferencemagazine.substack.com/p/the-masters-of-our-destiny</link><guid isPermaLink="false">https://inferencemagazine.substack.com/p/the-masters-of-our-destiny</guid><dc:creator><![CDATA[Jack Wiseman]]></dc:creator><pubDate>Tue, 11 Mar 2025 17:01:18 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa865596-111e-4e2e-b4ed-2f5da3fb6e04_1600x1130.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Dr Strangelove </em>is about the irrationality of mutually assured destruction. The presentation is a particularly American irrationality: General Ripper launches the pre-emptive strike because he thinks the communists have poisoned the water source to reduce the &#8216;purity&#8217; of the people. The only character who notices this mistake early in the plot is Lionel Mandrake, the <em>British </em>Air Attache<em>. </em>But he&#8217;s outside the command structure and there&#8217;s nothing he can do about it. At one point Mandrake (in the West End version) says to Ripper, &#8220;I <em>insist </em>that you give me the access codes at once&#8230;<em>please</em>&#8221;. It&#8217;s a broader metaphor for Britain&#8217;s self-perceived relationship to America: more sensible, detached, far-sighted, but less powerful. Mandrake is a stand-in for the British angst; however strongly he feels, however erroneous US policy is, it doesn&#8217;t matter. This week has laid this bare. The Leader of the Free World reminded the President of an invaded European country, &#8220;With us, you have the cards. Without us, you don&#8217;t have any cards.&#8221;<br><br>If Britain (and Europe) want to break this pattern, it&#8217;ll require first articulating a theory of technological sovereignty. What does national sovereignty depend upon? And how will AGI, and the acceleration of science and technologies which that will enable, change the answer to this question?</p><div><hr></div><p>The key question of the last week has been to what extent Europe can backstop Ukraine, as the US pause their involvement. The US commitment to any negotiated settlement is uncertain: perhaps they will provide a <em>de facto</em> security guarantee, through a mineral deal, but would this hold off a Russian invasion? Perhaps they provide a <em>de jure </em>security guarantee, but it is unclear if they would be committed, if this was tested again. The incoming Undersecretary for Defence, Elbridge Colby, wrote in his 2021 book, <em>The Strategy of Denial</em>:</p><blockquote><p>[T]he United States might very well not fill the gap in Eastern NATO left by any European unwillingness to strengthen their own defense efforts. Indeed, my argument in this book is that the United States <em>should not</em> plug these gaps. If China succeeds in its focused and sequential strategy in Asia, it can establish hegemony over the world's most important region. If Russia succeeds in a fait accompli in Eastern Europe, it will call NATO into question and open the East to Moscow's predominance, but it will not be able to dominate the wealthiest parts of the continent.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p></blockquote><p>He could not be clearer about their intentions here!</p><p>Without either of these US guarantees, to what extent would a 20,000-strong European peacekeeping force in Ukraine be respected? When the French, Germans, and Ukrainians negotiated the Minsk Accords in 2015; Russia later rescinded. If a settlement were to fail, to what extent would Europe be able to make up the shortfall in US support? </p><p>The EU and the UK could find <em>the money</em> if they had to. Together, they have an annual GDP of more than 20 trillion euros; while over three years, the US Congress had appropriated $175 billion for Ukraine and provided $65.9 billion in military support.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!3Eo1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd023c67-fd6d-46dc-98d3-6ad80085275d_1432x1074.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!3Eo1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd023c67-fd6d-46dc-98d3-6ad80085275d_1432x1074.png 424w, https://substackcdn.com/image/fetch/$s_!3Eo1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd023c67-fd6d-46dc-98d3-6ad80085275d_1432x1074.png 848w, https://substackcdn.com/image/fetch/$s_!3Eo1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd023c67-fd6d-46dc-98d3-6ad80085275d_1432x1074.png 1272w, https://substackcdn.com/image/fetch/$s_!3Eo1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd023c67-fd6d-46dc-98d3-6ad80085275d_1432x1074.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!3Eo1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd023c67-fd6d-46dc-98d3-6ad80085275d_1432x1074.png" width="1432" height="1074" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fd023c67-fd6d-46dc-98d3-6ad80085275d_1432x1074.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1074,&quot;width&quot;:1432,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!3Eo1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd023c67-fd6d-46dc-98d3-6ad80085275d_1432x1074.png 424w, https://substackcdn.com/image/fetch/$s_!3Eo1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd023c67-fd6d-46dc-98d3-6ad80085275d_1432x1074.png 848w, https://substackcdn.com/image/fetch/$s_!3Eo1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd023c67-fd6d-46dc-98d3-6ad80085275d_1432x1074.png 1272w, https://substackcdn.com/image/fetch/$s_!3Eo1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd023c67-fd6d-46dc-98d3-6ad80085275d_1432x1074.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Source: <a href="https://www.ft.com/content/b4393738-c7da-47ca-82db-99c2c052afec">Financial Times</a></figcaption></figure></div><p>One has to wonder, <strong>why is there such a weak exchange rate between money and sovereignty? </strong>Why does European backing seem so weak by comparison?</p><p>The crucial difference is that Europe does not have the same state-of-the-art capabilities to provide Ukraine. The nature of wars is changing. Either you need &#8220;cheap mass&#8221;, lots of inexpensive drones, for example, or exquisite capabilities. Ukraine&#8217;s drone manufacturing is larger than any other European country and the US was providing its state-of-the-art capabilities.</p><p>Second-rate European capabilities are a poor substitute for the very best. American Patriot Missiles have a longer range than alternative European air defences and can neutralise faster-moving missiles. Likewise, American counter-battery artillery has a longer range, and is actually produced at scale. American electronic warfare offers more generalised drone and precision missile jamming, whereas European countries can only offer point-solutions. American intelligence, recently paused, offers real-time visibility of attacks, where Britain can only offer lower latency. Starlink continues to run, though if Elon were to turn it off, the alternatives are significantly worse. Starlink has 7,000 satellites, <a href="https://www.bloomberg.com/news/articles/2025-03-07/ukraine-s-dependence-on-starlink-in-war-won-t-be-easy-to-break">whereas the European replacement has just 600.</a> In sum, if Ukraine continues to fight beyond the next couple of months it will do so with patchier, shallower, and lower-scale defences.</p><p>In this case, Ukraine&#8217;s sovereignty rests on deep supply chains for &#8220;cheap mass&#8221; and guaranteed access to the very best capabilities. Without which, it has no cards.</p><p>How will this change in the future?</p><div><hr></div><h3>The &#8220;Compressed 21st Century&#8221;</h3><p>The most important change to national power will be the development of powerful AI systems.</p><p>In the most aggressive view, Dario Amodei, the CEO of Anthropic, <a href="https://darioamodei.com/machines-of-loving-grace">has written</a> that AI systems with the cognitive capabilities of a Nobel Prize-level scientist in all domains could be created &#8220;as early as 2026, though there are also ways it could take much longer&#8221;. In some domains, he thinks this could lead to a compressed 21st century&#8212;100 years of progress in just a decade. The Chief Scientist at Meta has expressed the most sceptical view of any lab leader: he thinks that human-level AI could take a decade. However, Mark Zuckerberg has also said that AI systems will be able to <a href="https://www.businessinsider.com/mark-zuckerberg-meta-ai-replace-engineers-coders-joe-rogan-podcast-2025-1">perform the work of a &#8220;mid-level software engineer at Meta&#8221;</a> by the end of 2025. We should be preparing for very fast progress.</p><p>Already, the public state of the art already <a href="https://metr.org/blog/2024-11-22-evaluating-r-d-capabilities-of-llms/">outperforms human ML engineers on some tasks</a>, <a href="https://cdn.openai.com/deep-research-system-card.pdf">could have written 42% of OpenAI&#8217;s changes to their code base</a>, scores comparably to PhD-level experts on tests of scientific expertise. For introductions to technical AI progress, see <a href="https://inferencemagazine.substack.com/p/agi-is-an-engineering-problem">&#8220;AGI is an engineering problem&#8221;</a> and <a href="https://inferencemagazine.substack.com/p/on-o1">&#8220;on o1&#8221;</a>. Crucially, even if AI progress plateaued at human-level, it would be an enormously important tool. Some people have speculated that it will be possible to run millions of copies, much faster than humans can process information.</p><p>The most critical step is <em>what comes after human-level AI</em>. When AI systems could automate all the steps of the AI research and development process, including re-training improved copies of themselves, there could be a very fast acceleration in AI capabilities. This period of recursive self-improving has been termed an &#8220;Intelligence Explosion&#8221;. In our view, <a href="https://inferencemagazine.substack.com/p/how-much-economic-growth-from-ai">this will be bottlenecked on the most aggressive time horizons</a> (~2 years), but it is possible in the future.</p><div><hr></div><h3>How does powerful AI affect national power?</h3><p>Some researchers and AI lab leaders have written that whoever reaches the Intelligence Explosion first might be able to parlay this lead into a decisive strategic advantage over all other countries.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> The thinking goes that this advantage could be used to create a unipolar world order, negotiated or otherwise. This high-level abstraction is useful to keep in mind, but in a more concrete manner there are three ways AI will change national power.</p><p>First, AI is dual-use. It can be turned into a weapon much more easily than previous general-purpose technologies, like electricity or computers. In <a href="https://www.nationalsecurity.ai/">a recent paper</a>, Eric Schmidt, the former Google CEO, and his coauthors suggest AI cyberweapons would be able to &#8220;suddenly and comprehensively destroy a state&#8217;s critical infrastructure&#8221;. AI systems could also be used in drone jamming, targeting, and stealth capabilities.</p><p>Second, just as AI systems will be able to automate all steps of the AI research process, it will also be able to augment or take-over other R&amp;D processes. Think: drones, robots, sensors, chips, missiles. <a href="https://situational-awareness.ai/wp-content/uploads/2024/06/situationalawareness.pdf">An essay by a former OpenAI researcher</a> summarised this:</p><blockquote><p>Imagine if we had gone through the military technological developments of the 20th century in less than a decade. We&#8217;d have gone from horses and rifles and trenches, to modern tank armies, in a couple years; to armadas of supersonic fighter planes and nuclear weapons and ICBMs a couple years after that; to stealth and precision that can knock out an enemy before they even know you&#8217;re there another couple years after that.</p><p>That is the situation we will face with the advent of superintelligence: the military technological advances of a century compressed to less than a decade.</p></blockquote><p>For this reason, Eric Schmidt&#8217;s paper also suggests that some AI &#8220;superweapons&#8221; could obfuscate mutually assured destruction, which keeps the nuclear balance in check. AI could be used to create a &#8220;transparent ocean&#8221;, that means submarines can no longer operate in stealth; it could enable a nuclear power to find its adversary&#8217;s land nuclear launchers, or deceive its adversary about its intention or capabilities. The delicate equilibrium currently depends on a robust escalation ladder, which AI systems could shake.</p><p>Third, AI will boost productivity across almost all industries. A recent book, <em><a href="https://www.amazon.co.uk/Technology-Rise-Great-Powers-International/dp/0691260346">Technology and the Rise of Great Powers</a></em>, Jeffrey Ding makes the case that national power shifts in previous Industrial Revolutions are the result of deep, broad deployment across many sectors, rather than arising from the eureka moment of discovery. We have <a href="https://inferencemagazine.substack.com/p/how-much-economic-growth-from-ai">written previously</a> about the rate of deployment we expect through R&amp;D and the cognitive economy. The economic advantage from AI could be more important in the short-term, as many of the military applications of AI depend on very capable systems. Over time though, differential adoption and productivity would have an exponential effect on any country&#8217;s economic power. (It is important to note that an Intelligence Explosion would reduce the relative importance of this factor, however.)</p><p>Whether the most dangerous capabilities are unlocked in two years or ten, the path is clear: AI will be totally essential for military and economic power.</p><div><hr></div><h3>What does this mean for the world order?</h3><p>In a far-sighted essay from 2018, <a href="https://www.ianhogarth.com/blog/2018/6/13/ai-nationalism">AI nationalism</a>, Ian Hogarth predicted the emergence of &#8220;a kind of dependency would be tantamount to a new kind of colonialism&#8221;, whereby the world is split into countries <em>without</em> frontier AI capabilities, who are forced to depend economically and militarily on countries who do. This is sometimes summarised as being an &#8220;AI taker&#8221; or an &#8220;AI maker&#8221;. Such thinking was based on the work of Kai Fu-Lee, who wrote in his book <a href="https://www.amazon.co.uk/AI-Superpowers-China-Silicon-Valley/dp/132854639X">AI Superpowers</a><em> </em>in 2019:</p><blockquote><p>I fear this ever-growing economic divide will force poor countries into a state of near-total dependence and subservience. Their governments may try to negotiate with the superpower that supplies their AI technology, trading market and data access for guarantees of economic aid for their population. Whatever bargain is struck, it will not be one based on agency or equality between nations.</p></blockquote><p>At present, capabilities seem to be more greatly diffused than the kind of &#8216;superintelligence-in-a-bottle&#8217; which Ian Hogarth and Kai Fu-Lee seem to have in mind. However, this currently depends on AI labs near the frontier continuing to make their best capabilities available, whether open-source or through the API. As <a href="https://inferencemagazine.substack.com/p/what-did-you-do-this-weekin-ai-research">we have written about previously</a>, it seems probable to imagine that the gap between the actual frontier and what AI labs make available to the public will grow with capabilities.</p><p>While the UK self-styles as an &#8216;AI superpower&#8217;, or at least <em>wanting to be an AI superpower</em>, there are no UK companies with state-of-the-art capability in any major step of the production general-purpose AI. (This would mean capability in energy, chip manufacturing equipment, chip fabrication, AI accelerator design, grid connection, gigawatt-scale datacentre capacity, datacom and telecommunications.) On what basis would the UK negotiate its access to frontier capabilities? </p><p>It could look something (slightly) like this:</p><blockquote><p>[Enter scene. The US President and staff, with AI labs, are sat across from the UK Prime Minister and staff.]<br><br>The US President kicks off: &#8220;We&#8217;d like to make a deal for your access to our frontier capabilities. For too long America has been taken advantage of by its allies. Would you be able to give us some additional training capacity for our AI labs?&#8221; If <a href="https://www.theverge.com/news/619063/uk-newspapers-covers-protest-government-ai-rights-proposal">the lobbying in 2025 was successful</a>, the Prime Minister would be forced to say, &#8220;Unfortunately not, Mr President, we decided to make it illegal to train models under our copyright rules.&#8221;</p><p>The President: &#8220;Not to worry, our American companies will continue to train their models on the work of UK creatives in the US instead, it matters not. Do you have any datacentre capacity <em>for inference </em>they might be able to use instead?&#8221;. Again the Prime Minister would be forced to respond: &#8220;Alas, it's &#8216;no&#8217; again I&#8217;m afraid. When we were deciding to build datacentres we blocked their construction <a href="https://www.cityam.com/deranged-government-blocks-data-centre-build-next-to-m25-in-case-it-ruins-the-green-belt/">to preserve the view from motorway bridges</a> nearby. However, we can offer you a large population of <a href="https://www.ft.com/content/81008bda-b6d1-4870-b43a-aa308485f313">rare bats</a> if you need to repopulate places where you built datacentres.&#8221;</p><p>&#8220;That&#8217;s a shame, Prime Minister, I saw you announced <a href="https://www.gov.uk/government/publications/ai-growth-zones-expression-of-interest/ai-growth-zones-submit-an-expression-of-interest">reforms to improve planning for datacentres</a>, if not completed datacentres can you offer us your future capacity?&#8221;</p><p>&#8220;Mr President, you must understand that in 2025, our grid people <a href="https://x.com/inferencemag/status/1882116776169009179">said they were</a> &#8216;very confident that we can accommodate the increasing power demand that would come from AI&#8217;, so unlike you, <a href="https://www.datacenterdynamics.com/en/news/trump-we-need-double-the-energy-we-currently-have-in-the-us-for-ai-promises-emergency-declaration-for-more-power/">we did not double our grid</a>.&#8221;</p><p>Exasperated, the President responds, &#8220;In 2025, <a href="https://www.rand.org/pubs/research_reports/RRA3572-1.html">the projections were showing</a> that AI accelerator orders in 2030 could require 300 gigawatts globally, what did you think was going to happen?&#8221;. The President sighs, and moves on, &#8220;I am told that getting a grid connection in the UK is falling from 10 years to 8 years, is there any chance we could at least have a grid connection in a few years?&#8221;.</p><p>&#8220;Ah, again, unfortunately, the only reason the grid connection queue is falling is because our <a href="https://www.theguardian.com/business/2025/jan/15/great-britain-energy-system-operator-blocks-access-grid-connection-queue">national operator has banned entering the queue</a>.&#8221;</p><p>The President: &#8220;Do you have any industrial manufacturing capacity at all; either for chips or for robots?&#8221;</p><p>&#8220;Ah, again, Mr President, we have the highest industrial energy prices in the world and we chose to become a &#8216;high-skill, high-wage&#8217; economy that doesn&#8217;t focus on low-value added tasks like manufacturing. However, we did become a clean energy superpower and our economy is focused on high value-added tasks like making films and doing financial services. Do you have any use for these things?&#8221;</p><p>&#8220;Well Prime Minister, with the US models that we&#8217;ve trained on the <em>entire corpus of British films, </em>so we can now sell back to you, the <em>ideal </em>British film. And our models are already extremely good at augmenting financial services in New York, so we expect London to become less important for us over time.&#8221;</p><p>&#8220;What can we offer you then?&#8221;</p><p>The President pauses, and looks up for a minute, takes a short breath, and says, &#8220;It would be great for American tourists who are rich from the AI wealth to be able to land more often at Heathrow. Anything you can do here?&#8221;<br><br><em>[End scene. Author&#8217;s note: Some artistic license was taken for effect. Also, some readers may note that Google DeepMind is based in London, but since it is a US company this does not seem to provide any support, and Arm designs a chip for each NVIDIA H100 server but it only handles non-core tasks like system management, so it seems reasonable to imagine there is no strategic benefit.]</em></p></blockquote><div><hr></div><h3>Sovereignty is a market failure.</h3><p>To begin to find a solution, it is first appropriate to look back to answer, how did the UK become so dependent? In 1962, two years before Stanley Kubrick created the ineffectual Lionel Mandrake, the former US Secretary of State commented that, &#8220;Great Britain has lost an empire and has not yet found a role&#8221;. This question was never really answered; the UK just followed the US course on neoliberalism. In effect, it was left to intellectuals at the University of Chicago and Mont Pelerin Society.</p><p>In the neoliberal conception, values and beliefs remain in the private sphere, and in the public sphere, there is just a minimal state to uphold the market. The big question, of what we value collectively, was left to the invisible hand. As Thatcher put it, &#8220;There is no such thing as society.&#8221; Just as in AI research we pick the objective and hillclimb towards that.</p><p>The UK has done this to the extreme. In investing terms, the UK took on very high factor exposure to globalisation: becoming an exporter of services and making fewer and fewer things.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Db7-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa865596-111e-4e2e-b4ed-2f5da3fb6e04_1600x1130.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Db7-!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa865596-111e-4e2e-b4ed-2f5da3fb6e04_1600x1130.png 424w, https://substackcdn.com/image/fetch/$s_!Db7-!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa865596-111e-4e2e-b4ed-2f5da3fb6e04_1600x1130.png 848w, https://substackcdn.com/image/fetch/$s_!Db7-!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa865596-111e-4e2e-b4ed-2f5da3fb6e04_1600x1130.png 1272w, https://substackcdn.com/image/fetch/$s_!Db7-!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa865596-111e-4e2e-b4ed-2f5da3fb6e04_1600x1130.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Db7-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa865596-111e-4e2e-b4ed-2f5da3fb6e04_1600x1130.png" width="724" height="511.325" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fa865596-111e-4e2e-b4ed-2f5da3fb6e04_1600x1130.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1130,&quot;width&quot;:1600,&quot;resizeWidth&quot;:724,&quot;bytes&quot;:191437,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Db7-!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa865596-111e-4e2e-b4ed-2f5da3fb6e04_1600x1130.png 424w, https://substackcdn.com/image/fetch/$s_!Db7-!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa865596-111e-4e2e-b4ed-2f5da3fb6e04_1600x1130.png 848w, https://substackcdn.com/image/fetch/$s_!Db7-!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa865596-111e-4e2e-b4ed-2f5da3fb6e04_1600x1130.png 1272w, https://substackcdn.com/image/fetch/$s_!Db7-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa865596-111e-4e2e-b4ed-2f5da3fb6e04_1600x1130.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://ourworldindata.org/grapher/electricity-generation">Total Electricity Generation</a>, Our World in Data</figcaption></figure></div><p>During the supply chain crunch in 2021, Ryan Petersen <a href="https://x.com/typesfast/status/1453753924960219145">wrote</a> that the issues were caused by an obsessive focus with return on equity:</p><blockquote><p>&#8220;To show great ROE almost every CEO stripped their company of all but the bare minimum of assets. Just in time everything. No excess capacity. No strategic reserves. No cash on the balance sheet. Minimal R&amp;D. We stripped the shock absorbers out of the economy in pursuit of better short term metrics.&#8221;</p></blockquote><p>Britain has &#8220;done a Boeing&#8221;: outsourced its supply chain, and forgot how to make things. Now the plane is falling apart as we are flying. In 2008, the UK was richer than the US per head, now the UK is poorer than all but the poorest US state. The North of England has become <a href="https://tomforth.co.uk/whynorthenglandispoor/">even poorer than former communist countries</a>, like East Germany and Poland. We eked out the gains of financialisation, but we didn&#8217;t make anything new in the real world. It turns out that <a href="https://employamerica.medium.com/a-brief-history-of-semiconductors-how-the-us-cut-costs-and-lost-the-leading-edge-c21b96707cd2">a lot of value exists in the connective tissue between steps in the supply chain</a>, because when you understand the whole process you can innovate. This is how SpaceX and Tesla have done so well.</p><p>Emmanuel Macron <a href="https://www.economist.com/europe/2024/05/02/emmanuel-macron-in-his-own-words-english">described the error of the neoliberal consensus in 2019</a>, which applies equally to Britain:</p><blockquote><p>&#8220;Europe has forgotten that it is a community, by increasingly thinking of itself as a market, with expansion as its end purpose. This is a fundamental mistake, because it has reduced the political scope of its project, essentially since the 1990s. A market is not a community. A community is stronger: it has notions of solidarity, of convergence, which we&#8217;ve lost, and of political thought."</p></blockquote><p>Hollowing out your industries, in pursuit of better GAAP metrics for quarter-end, is not just a bad economic decision, it is a spiritual hollowing out. There is no longer a political project or direction or values; we are &#8220;just individuals&#8221; in a fragile, exposed, competitive, global economy. Clearly this is not <em>all there is</em>. And for whatever &#8216;else&#8217; might be, sovereignty is a necessary precondition. Sovereignty is not priced by the market so it cannot be valued by the market alone.</p><div><hr></div><h3>Sovereignty, to do what?</h3><p>In some sense, being sovereign is intrinsically good. Even if an AI system could run the world &#8220;more optimally&#8221; or exactly the same as humans would, it would be a disappointing outcome. The option value; the freedom to choose otherwise, is worthwhile. But aside from this, it can be useful to reflect on <em>to what end </em>this will be valuable, when we think about why it is worth upholding.</p><p>One reason that Britain might have struggled to find a role in the second half of the 20th century, as Acheson pointed out, is that there is not <em>clearly </em>a &#8220;British project&#8221; in the same way there is an American experiment. The United States&#8217; founding was explicitly a project in self-government based on democracy, individual liberty, and the rule of law; in opposition to what it viewed as the tyranny of the Old World. Its self-conception as &#8220;the last best hope of earth&#8221; is both a useful fallback and self-corrective. The same sense of purpose, or direction, can be found in Britain too; if motivated as a contrast&#8230;</p><p>Given the UK&#8217;s weak position the economically optimal thing to do would be to become the 51st state&#8212;if the US would accept it&#8212;But if any politician suggested joining, there would probably be a revolt. One just has to look at the response in Canada to the Trump Administration&#8217;s suggestion that they might join the Union. Just this week, in Mark Carney&#8217;s first address as Canadian Prime Minister he said: <a href="https://www.bbc.co.uk/news/videos/czjep2ddynro">&#8220;Canada will never, ever be part of America&#8221;</a>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!cFtt!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e58fee5-de00-479e-9048-1641c8681ff7_1594x1590.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!cFtt!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e58fee5-de00-479e-9048-1641c8681ff7_1594x1590.png 424w, https://substackcdn.com/image/fetch/$s_!cFtt!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e58fee5-de00-479e-9048-1641c8681ff7_1594x1590.png 848w, https://substackcdn.com/image/fetch/$s_!cFtt!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e58fee5-de00-479e-9048-1641c8681ff7_1594x1590.png 1272w, https://substackcdn.com/image/fetch/$s_!cFtt!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e58fee5-de00-479e-9048-1641c8681ff7_1594x1590.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!cFtt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e58fee5-de00-479e-9048-1641c8681ff7_1594x1590.png" width="1594" height="1590" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9e58fee5-de00-479e-9048-1641c8681ff7_1594x1590.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1590,&quot;width&quot;:1594,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:707530,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!cFtt!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e58fee5-de00-479e-9048-1641c8681ff7_1594x1590.png 424w, https://substackcdn.com/image/fetch/$s_!cFtt!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e58fee5-de00-479e-9048-1641c8681ff7_1594x1590.png 848w, https://substackcdn.com/image/fetch/$s_!cFtt!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e58fee5-de00-479e-9048-1641c8681ff7_1594x1590.png 1272w, https://substackcdn.com/image/fetch/$s_!cFtt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9e58fee5-de00-479e-9048-1641c8681ff7_1594x1590.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>What explains this strong reaction, especially when there are so many advantages to joining in economic terms?</p><p>The most compelling explanation is that Britain, and others, have a slightly different flavour of the Western project, despite sharing a lot with the American cousins. To make two observations about the distinctiveness of Britain&#8230;</p><p>First, it has incredible longevity. &#8216;England&#8217; has been a nation for over a millennium. Only Denmark and Japan can make comparable claims. From this, comes a steadier, more rooted culture. Perhaps the aristocracy, and their focus on lineage and preservation of tradition, were the original longtermists! This is combined with a Whiggish consensus for improvement. The economic historian Anton Howes found the Industrial Revolution happened in Britain, not elsewhere, because of an &#8220;improving mentality&#8221;. Joel Mokyr wrote, too, that British workers wanted to accumulate &#8216;useful knowledge&#8217; and experiment pragmatically. In investing terms, buying Britain is buying a compounder: half a percent productivity growth a year, over centuries adds up. (It is <a href="https://www.thediff.co/archive/how-would-you-run-a-10000-year-endowment/">possible to tolerate large drawdowns in the long run</a>.)</p><p>Second, across this long period, the people have been unusually immune to extremism. Some have suggested this is because the elite are unusually responsive to the people. In George Orwell&#8217;s essay, <em>The Lion and the Unicorn,</em> he wrote, &#8220;The nation is bound together by an invisible chain&#8230;let popular opinion really make itself heard, let them get a tug from below that they cannot avoid feeling, and it is difficult for them not to respond.&#8221; While not strictly <em>popular </em>opinion, the abdication of Richard II, the restoration of the monarchy, and the Glorious Revolution are both unusual cases of a leader giving up their power in response to elite views. Likewise, Robert Tombs noted in <em>The English and Their History, </em>&#8220;It is hard to think of any major improvement in England since Magna Carta [1215] brought about by violence&#8230;[m]any of the things we consider pillars of liberty &#8212; the common law, trial by jury, habeas corpus, religious toleration &#8212; came not from popular protest but from politics of the Crown developed by royal judges.&#8221;<br><br>Corruption, by any international standards, is minor. There was a &#8216;scandal&#8217; when the Prime Minister received suits. Another Prime Minister was criticised for redecorating Downing Street. While it might have gone on too long after the COVID lockdown parties were revealed, there was eventually a cascade of resignations by Conservative ministers and the leader was replaced. The system of informal principles worked.</p><p>The compressed 21st century is likely to be an enormously turbulent period. When I think about the things that could go wrong&#8212;an impetus to use models which could be misaligned, power grabs, international conflict, enormous inequality, or gradual disempowerment&#8212;it seems clear that the UK has something to offer. That is, to bring to bear its flavour of the Enlightenment project on the development and governance of AI. To be the harbinger for reasonableness, patience, common sense, with whiggish eagerness for improvement; to complement the American frontier, day I say cowboy, spirit. It is extremely serious that we get this right: <a href="https://www.businessinsider.com/elon-musk-20-percent-chance-ai-destroys-humanity-2024-3">Elon Musk</a> and <a href="https://www.theguardian.com/technology/2024/dec/27/godfather-of-ai-raises-odds-of-the-technology-wiping-out-humanity-over-next-30-years">Geoffrey Hinton</a> have both said there is a 20% chance that AI kills us all.</p><p>The UK has a fair-minded tradition of scientific inquiry, has made public goods available for the world before&#8212;like common law, the joint stock corporation, and the parliamentary system&#8212;as AI should be, and has a different emphasis to America, which is worth having too. Who else will project the spirit of Locke, Hume, and Mill into the lightcone of the universe?<br><br>That, or there are two other options: join as the 51st state, or become a cold, wet version of Portugal.</p><div><hr></div><h3>Making technology to uphold sovereignty</h3><p>Just as the US has used the dollar as a tool of statecraft, so too will countries use state-of-the-art capabilities as a foreign policy tool. The US was able to change the ruler of Iran <a href="https://www.chinatalk.media/p/american-power-in-the-age-of-economic">leveraging international banks&#8217; access to dollars</a>, and perhaps the war in Ukraine will be &#8220;switched off&#8221; by the withdrawal of US capabilities. In the future, if you run someone else&#8217;s models, on someone else&#8217;s servers, made using their tools; you are not in control. As Sam Currie highlights in <a href="https://curriesam.substack.com/p/the-future-of-britains-economic-statecraft">his excellent recent piece</a>, during the Pandemic, the US attempted to seize all Moderna vaccines and diagnostic supplies that were manufactured in the US. Only when Germany said it would withhold access to reagents from their domestic firms, was this avoided.</p><p>From this, the goal is clear. A country upholds its technological sovereignty <em>not </em>by trying to domestically produce everything&#8212;this would just lead to subpar capabilities&#8212;but having strategic leverage (state-of-the-art) in some areas, to guarantee access to all necessary capabilities on good terms.</p><p>What are the necessary capabilities? A paper by Jeffrey Ding and Allan Dafoe provides a framework for determining <a href="https://arxiv.org/pdf/2001.03246">the logic of strategic assets</a>. In their rubric, there are three features of a technology which determine its importance: how valuable it is economically or militarily, to what extent it creates benefits or costs that companies don&#8217;t capture (and so would be underinvested in), and to what extent the benefits or costs can be &#8216;nationalised&#8217; by the country where it is produced. There are three &#8216;logics&#8217; which amplify the strategic importance further. The <em>cumulative logic</em>; whether initial advantages grow over time, <em>infrastructure logic</em>; whether it supports many technologies or sectors, <em>dependency logic</em>; whether it is at risk from concentrated supply or potential disruption.</p><p>This is why the foundation model layer is so important. It ticks all the boxes for importance. Foundation models will be a central input into all future frontier science and technology progress, almost all processes with a cognitive element, and will have military applications. The benefits of general-purpose technologies spread far throughout the economy. While countries cannot &#8216;nationalise&#8217; open-weight models which have already been released, labs can withdraw API access, impose usage limits, or not release models at all. Next, being early to develop foundation models has compounding returns: once the automation of AI R&amp;D has begun, it will be almost impossible to join in later. It will be like an &#8216;infrastructure layer&#8217; for cognitive work (&#8220;a steam train for the mind&#8221;), and the frontier is made in just two countries.</p><p>Beyond this layer, there are five questions of vital importance for all countries:</p><ol><li><p>Do you have abundant electrons?</p></li><li><p>Do you have abundant FLOP?</p></li><li><p>Do you have the most capable and abundant tokens?</p></li><li><p>Do you have the cheapest, and most capable, robots and drones?</p></li><li><p>Do you have the lowest latency and highest throughput communication networks?</p></li></ol><p>To simplify: energy, chips, models, robots, drones, and networking. How sure is the supply chain for each of these? On what terms is your supply guaranteed? We are all believers in the legalistic global order when the sun is shining. Let&#8217;s hope our counterparties are too, if the storm comes.</p><div><hr></div><h3>Conclusion</h3><p>While the overall tone of this essay has been to embrace issues of national power, sovereignty, and defence; this is not the impression I hope to leave. None of these instrumental goals are for their own sake. As I hope to have made the case for, my hope is that if the UK has sovereign capability, it is good not just for the UK, but as a counterweight to excess variance in the world. Things are dangerous now, and the development of powerful AI could make the next decade even more turbulent. A sovereign AI effort, I hope, could help to reduce race dynamics between great powers, and alter the emphasis from a potential arms race towards a scientific endeavour which would benefit all humanity.</p><p>Optimistically, AI sovereignty for Britain could be the lynchpin of a new pluralistic, tolerant, and peaceful world order.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>The Strategy of Denial (2021), Elbridge Colby, p.276</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Superintelligence (2014), Nick Bostrom; Situational Awareness (2023), Leopold Aschenbrenner.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[Peak Brussels]]></title><description><![CDATA[&#8220;Who do I call if I want to speak to Europe?&#8221; &#8212; Kissinger]]></description><link>https://inferencemagazine.substack.com/p/peak-brussels</link><guid isPermaLink="false">https://inferencemagazine.substack.com/p/peak-brussels</guid><dc:creator><![CDATA[Anton Leicht]]></dc:creator><pubDate>Sun, 02 Mar 2025 23:48:29 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c61df20-b545-4c7b-9acb-75940496383f_1010x1010.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<blockquote><p>&#8220;Who do I call if I want to speak to Europe?&#8221; &#8212; Kissinger</p></blockquote><p>It was sort of an accident that the EU was first to regulate AI. Issues will tend to drift up to the European level if they are politically uninteresting to national governments, and needing an unpleasant solution or technical implementation. In October 2020, AI seemed to fit the latter description and so the drafting process for regulation began.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><p>ChatGPT&#8217;s explosive growth interrupted this process. The earlier draft of the AI Act placed all of the regulatory burden on end-deployers of AI products, not anticipating the shift to foundation models, and so additional provisions for &#8216;General-Purpose AI&#8217; needed to be made. But the greater change was the shift in perception: AI was no longer an issue of low political salience. Once countries began to notice that foundation models would be the next general-purpose technology, negotiating the additional provisions became much more difficult. Germany, Italy, and France were concerned the AI Act would hamstring nascent foundation models providers and railed against regulation. A barebones proposal prevailed in the negotiations, with substantial implementation work left to do.</p><p>With hindsight, the introduction of the AI Act will be, we believe, the high point of the European Commission&#8217;s relative importance in AI policy. A set of unassailable macro forces will pull power away from Brussels:</p><ol><li><p>The models will get a lot better, quickly.</p></li><li><p>As this happens, access to powerful AI becomes increasingly important for national productive capacity and so EU member states will face mounting pressure to weaken the enforcement of the AI Act, and or make bilateral agreements with AI makers, to access the most advanced models.</p></li><li><p>Likewise, access to powerful AI becomes increasingly necessary for security, where too, member states will be minded to make bilateral agreements with AI makers for reliable access to state-of-the-art models.</p></li><li><p>In this context, the Trump administration <a href="https://www.youtube.com/watch?v=pCOsgfINdKg">has made clear</a> they will not tolerate overburdensome regulation of their tech companies by the EU.</p></li></ol><p>This combination of forces exacerbates existing headwinds: national governments have grown evermore sceptical of the EU&#8217;s approach to tech regulation&#8212;the Digital Services Act and GDPR are often blamed for the weakness of Europe&#8217;s digital economy.</p><p>The Commission has assembled a group of experts to set out a Code of Practice which will set out implementation of the AI Act to the most powerful models. But de facto authority for this process has spread beyond Brussels: national economic interest and transatlantic pressure limit its teeth, and foundation model providers can choose not to opt into the Code of Practice altogether. Down this path, they would face alternative case-by-case enforcement of the Act, but who is to say whether the Commission would have the political backing to take dissenters to court? Arriving at a strong, politically achievable code that is a <em>blueprint rather than a cautionary tale </em>is a very thin needle to thread. Perhaps the strongest influence of the AI Act is its influence on others; positively or negatively.</p><p>The next set of questions of AI policy will relate to the supply chain, encouraging adoption, and governing agents.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> On these issues, we can expect much higher political salience and so we can expect much stronger national engagement. There will be a looming threat of falling further behind, of labour market disruption, and countries will need to be able to redistribute between &#8216;winners&#8217; and &#8216;losers&#8217;. Historically-powerful labour unions in France and Germany will make demands through their national parties where they have a much stronger footprint. Service businesses will demand better access to inference compute and support for adoption initiatives. The EU is already perceived as having a weak track record on issues of competitiveness and supply chain buildup.</p><p>These next issues are likely to remain with the national government, as they will move faster in areas of clear national interest and local need. Either the EU policies will be dead on arrival in the council, or greatly influenced by existing national approaches: it is no sign of Brussels&#8217; influence if the EU parliament passes a law already on the French and German books. Even in the purportedly &#8216;European&#8217; approach at present&#8212;the Commission President&#8217;s announcement of &#8364;200 billion investment for AI infrastructure&#8212;comes from a combination of private funding, member state investment, and EU funds that have previously been restricted by member states; rather than any discretionary Commission funding.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> This is the kind of approach that the Commission President could be referring to with the idea of an EU-led &#8216;CERN for AI&#8217;. But wherever an EU megaproject might seem like evidence for a prominent role for Brussels, it often turns out that any major member state can set off a chain reaction to question its merits, demand local favouritism, or choke off its funding at will. Brussels is hardly in the driving seat.</p><p>As with economic policy, when the security and geopolitical implications of AI are sharpened, national governments will move to make deals with AI makers. Already, some European countries are being treated preferentially in the tiers of US export controls on frontier AI chips. In tier two countries, commercial orders of GPUs are capped at 50,000 per year. A small number of countries &#8212; France, Germany, Italy, the Netherlands, the Scandinavian countries, and perhaps Poland &#8212; are likely to be treated preferentially by the US for access to models. The incentive for any one of these actors to defect from the EU negotiating as a bloc will only get stronger as the speed of improvement hastens, and the dominance of the technology becomes clearer.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!U_4q!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26f8993f-792e-4a5d-ac20-3494addfafac_556x456.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!U_4q!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26f8993f-792e-4a5d-ac20-3494addfafac_556x456.png 424w, https://substackcdn.com/image/fetch/$s_!U_4q!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26f8993f-792e-4a5d-ac20-3494addfafac_556x456.png 848w, https://substackcdn.com/image/fetch/$s_!U_4q!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26f8993f-792e-4a5d-ac20-3494addfafac_556x456.png 1272w, https://substackcdn.com/image/fetch/$s_!U_4q!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26f8993f-792e-4a5d-ac20-3494addfafac_556x456.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!U_4q!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26f8993f-792e-4a5d-ac20-3494addfafac_556x456.png" width="556" height="456" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/26f8993f-792e-4a5d-ac20-3494addfafac_556x456.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:456,&quot;width&quot;:556,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!U_4q!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26f8993f-792e-4a5d-ac20-3494addfafac_556x456.png 424w, https://substackcdn.com/image/fetch/$s_!U_4q!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26f8993f-792e-4a5d-ac20-3494addfafac_556x456.png 848w, https://substackcdn.com/image/fetch/$s_!U_4q!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26f8993f-792e-4a5d-ac20-3494addfafac_556x456.png 1272w, https://substackcdn.com/image/fetch/$s_!U_4q!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F26f8993f-792e-4a5d-ac20-3494addfafac_556x456.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The US AI diffusion framework already makes a difference between EU members. <a href="https://www.rand.org/pubs/perspectives/PEA3776-1.html">RAND</a></figcaption></figure></div><p>This mirrors the joint European initiative to procure COVID vaccines. The EU remained in lockstep, because of the actions of Chancellor Merkel in particular, but the delayed and patchy vaccine rollouts cost the cause of collective action in the future. Received wisdom is that the EU&#8217;s most advanced economies paid the price for this. With the benefit of hindsight, a new security situation, and fewer Europhiles in the national governments; it seems hard to imagine an EU-led approach. EU leaders would need to commit to it unequivocally, and Brussels would need to prove itself worthy of that commitment.</p><p>So while at present the AI policy discourse has Brussels as a central actor, the transatlantic or inter-European political currents will pull away from this unstable equilibrium as AI gets more capable. If &#8212; for whatever reason &#8212; you want to dial Europe on AI policy in future, you might well have to call Paris, Berlin, and The Hague instead. Maybe Brussels will get to listen in.<br></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>GPT-3 was released in June 2020.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Long-horizon agents are not well covered by the AI Act.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>Euractiv has a <a href="https://www.euractiv.com/section/tech/news/breaking-down-europes-announced-e317-billion-for-ai/">full breakdown</a>.</p></div></div>]]></content:encoded></item><item><title><![CDATA[What did you do this week…in AI research?]]></title><description><![CDATA[(Don't worry, no need to respond in 5 bullet points or less.)]]></description><link>https://inferencemagazine.substack.com/p/what-did-you-do-this-weekin-ai-research</link><guid isPermaLink="false">https://inferencemagazine.substack.com/p/what-did-you-do-this-weekin-ai-research</guid><dc:creator><![CDATA[Jack Wiseman]]></dc:creator><pubDate>Sat, 01 Mar 2025 13:17:30 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!9o2z!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c61df20-b545-4c7b-9acb-75940496383f_1010x1010.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>It is popular, <a href="https://www.bloomberg.com/news/articles/2025-02-25/musk-warns-federal-workers-to-answer-next-email-or-be-fired">at least in some parts of the world</a>, to use very short surveys to take the pulse on organisational productivity. I don&#8217;t intend to make a value judgement on the use of, &#8220;What did you do this week?&#8221; in the federal government but I would like to propose that every three months, the AI labs could conduct a survey which asks about 40 researchers two questions:</p><ol><li><p>How much productivity uplift, compared with 2023, are you getting from AI systems <em>right now</em>?</p></li><li><p>How much productivity uplift, compared with 2023, do you expect to get from AI systems <em>in 6 months</em>?</p></li></ol><p>The researchers would answer with just a percentage<em> </em>for each question, and the results would be published. It would take just a minute!</p><h3>Why would this survey be useful?</h3><p>At a high level, three things are true:</p><ol><li><p>AI systems were <strong>previously unable to improve AI researcher productivity</strong>, but in the last few months;</p></li><li><p>AI systems <strong>are providing some noticeable benefit to AI researchers&#8217; output</strong>.</p></li><li><p>AI systems will be used to partially automate more steps in the research process before an AI system would be able to <a href="https://en.wikipedia.org/wiki/Recursive_self-improvement">&#8216;recursively self improve&#8217;</a>&#8212;by wholly automating the research process, and re-training improved copies of themselves.</p></li></ol><p>It would be very useful to be able to plot, over time, how much do researchers think that AI systems are giving them a productivity uplift, to be able to notice when we should expect to see extreme jumps in capability because of complete automation. (I realise it might be stating the obvious, but recursive self-improvement might lead to <em>very fast </em>jumps in AI capabilities, far beyond human level.)</p><p><strong>At the moment, we are practically &#8216;flying blind&#8217; about how soon superintelligence could come.</strong></p><p>Our current sources are anecdotes, press interviews, and essays from people at the labs, and model autonomy evaluations.</p><p>Without exhaustively listing examples of quotes from the lab leaders and researchers, here are some examples:</p><ul><li><p>Sam Altman has suggested we are <a href="https://ia.samaltman.com/">a few thousand days away from superintelligence</a>.</p></li><li><p>Dario Amodei has said that <a href="https://darioamodei.com/machines-of-loving-grace#basic-assumptions-and-framework">we could have a &#8216;Nobel Prize-level&#8217; scientist in all scientific domains</a> in as little as two or three years. (This could be used to automate research, to create superintelligence.)</p></li><li><p>A researcher from OpenAI tweeted that <a href="https://x.com/mcaleerstephen/status/1878555949662666895">&#8220;controlling superintelligence is a short term research agenda&#8221;</a>.</p></li></ul><p>Some people in the mainstream will dismiss comments like this on the grounds that AI labs need to fundraise or that Silicon Valley generally tends to &#8216;hype&#8217; emerging technologies. Irrespective of whether these critiques are correct, it seems the AI labs would be doing a disservice to ordinary people if they do not provide a clear grounding of that path, which could take the form, &#8220;9 months ago, our researchers thought they were getting a 25% output improvement from using AI systems, compared with being unaided, and now they believe overall, they are getting a 75% output improvement against 2023 benchmarks.&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> Conversely, it would be useful for those who are sceptical of very fast AI progress to substantiate how the productivity uplift that researchers are getting from AI models is flatlining, if in fact it is.</p><p>Public statements on progress are valuable, but are not <em>that rigorous</em>, and the absence of context does not aid in the world beginning to prepare for very capable AI systems.</p><p>Otherwise, there are <strong>two good public tests of model autonomy</strong>, but these are weak guides to estimate how useful the models actually are in real-world settings. <a href="https://arxiv.org/abs/2411.15114">RE-Bench</a> from METR tests the model&#8217;s ability to perform seven realistic, but self-contained ML engineering tasks and <a href="https://arxiv.org/abs/2410.07095">MLE-bench</a> uses 75 ML engineering tasks from Kaggle, a platform for doing online coding competitions. This is useful insofar as it allows us to understand how the models perform on end-to-end tasks of medium-length (hours) but it doesn&#8217;t capture: where is it actually rational to deploy models in the real world, and how useful is this actually for these jobs? It feels difficult to say anything beyond: &#8220;The models are quite useful, if a little unreliable, for narrow tasks like catching bugs, code autocomplete, and optimising kernels for a given architecture, where it makes sense to do integration work.&#8221;</p><p>As we move forwards, evaluations will be even more difficult:</p><p>Public or pre-deployment evaluations cannot capture the productivity uplift from models which are only deployed internally. As models become more powerful, it is reasonable to imagine AI labs will give their researchers access to use models for longer, before sharing with the outside world, in order to ensure models are safe and to improve their productivity differently. Evaluations will be unable to provide any indication of what kind of capability, or potential advantage, these researchers are getting.</p><p>Even then, it will be more challenging to get human controls for long horizon evaluations. We need to compare the model&#8217;s performance to a human baseline, ideally using lab researchers for the most representative test. The current human baselines are taken for time increments from 2 to 64 hours, but as the length of the task we evaluate model performance gets even longer, this gets more difficult.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> Imagine saying, &#8220;We&#8217;d like some of the world&#8217;s best researchers to come and solve these test problems for a week to compare them to model baselines,&#8221; clearly the labs are too busy for this! To account for this, METR are planning to do <a href="https://x.com/METR_Evals/status/1894257205680967907">open source developer uplift evaluations</a> and OpenAI have shown that Deep Research could make 42% of the pull requests (code edits) in their codebase.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a></p><p>Asking the researchers for their subjective impression is one way to mitigate this.</p><h3>Why would this survey be challenging?</h3><p>This is not a silver bullet, by any stretch! There are a number of reasons this survey has feasibility challenges or could have weaker explanatory power:</p><ul><li><p>In general, it is bad to burden the researchers with surveys! Very few stories of brilliant research environments involve lots of interruptions from people with clipboards. It is sensible to be wary of a &#8216;slippery slope&#8217; whereby each marginal question feels reasonable to add, but then researchers end up spending half their day filling out forms. However, on balance, it is also the case that very few research environments have tried to build superintelligence, so asking about 40 people, every three months, to complete a form that will take literally a minute feels proportionate.</p></li><li><p>It is possible that researchers&#8217; perceptions of their productivity uplift do not reflect their actual usefulness. On balance, it seems worthwhile nonetheless: researchers will often &#8216;vibe check&#8217; models, so even if it is an aggregation of their vibe checks on the usefulness of AI systems it still provides some indicator. If there is a systemic bias, the trendline will be valuable, even if the absolute values are not.</p></li><li><p>Finally, it is possible that there will be incentives for AI labs to encourage researchers to under- or over-state the productivity uplift they are getting. It seems good not to be too cynical in this regard, and only put so much emphasis on this datapoint. Perhaps this concern could be mitigated if the survey was conducted by a trusted third party &#8212; like the AI Security Institute, Epoch AI, or other evaluators &#8212; and partially anonymising the results.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a></p></li></ul><p>To step back, I expect almost everyone would agree that in the ideal case, superintelligence should be built in a maximally transparent way, but given the current equilibrium, this also needs to be achieved without compromising commercial or national interests. A two-question survey would be a low-cost and high-value step towards greater openness.</p><p></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Numbers are illustrative, I do not think anyone is getting a 75% productivity uplift yet.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>This is particularly true as the highest level of talent will more strongly differentiate on the longest-horizons.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p><a href="https://cdn.openai.com/deep-research-system-card.pdf">Deep Research System Card</a>, OpenAI, February 2025, p.33</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>The way I would imagine this is that the answers are published not by naming each lab and listing the scores, but rather here is the average at the &#8216;leading lab&#8217; and the &#8216;industry average&#8217; across all respondents.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[How the UK could build 3 gigawatts of new nuclear power by July 2027]]></title><description><![CDATA[Building new nuclear power in the UK is deeply broken&#8212;it is slow to approve; too slow to build, too expensive, and all too often asks for state subsidy.]]></description><link>https://inferencemagazine.substack.com/p/how-the-uk-could-build-3-gigawatts</link><guid isPermaLink="false">https://inferencemagazine.substack.com/p/how-the-uk-could-build-3-gigawatts</guid><dc:creator><![CDATA[Jack Wiseman]]></dc:creator><pubDate>Mon, 17 Feb 2025 01:24:15 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!9o2z!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c61df20-b545-4c7b-9acb-75940496383f_1010x1010.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Building new nuclear power in the UK is deeply broken&#8212;it is slow to approve; too slow to build, too expensive, and all too often asks for state subsidy. This is not a law of physics: other countries build reactors many times more quickly and cheaply, and the UK used to be able to. Since the UK last added a nuclear reactor, it has turned off 26 reactors and the rest of the world has added 148.</p><p>The Government has <a href="https://www.gov.uk/government/news/government-rips-up-rules-to-fire-up-nuclear-power">recognised there is a problem</a>, and announced their intention to &#8220;rip up the rules to fire up nuclear power&#8221;. To build nuclear power as quickly as possible; smaller is better. <strong>Giga-scale reactors </strong>would take a minimum of 5 years, <strong>Small Modular Reactors (</strong>roughly 500 to 100 megawatts) take 2 to 5 years, but <strong>Micro Modular Reactors</strong> (from<strong> </strong>1 to 50 megawatts) can be built in less than 2 years. I&#8217;ve spoken to developers who say this would be possible, but others have expressed scepticism. Setting this aside and being charitable to the companies; <strong>what would the Government need to do to make this possible?</strong></p><p>Working backwards, the fabrication of micro nuclear reactors would need to begin at the end of this year; and so the three pillars of nuclear approvals &#8212; licensing, permitting, and planning &#8212; would need to be reformed to allow this. This would probably require primary legislation within the next three months.</p><h3>What would that primary legislation have to do?</h3><ol><li><p>Create a regulatory sandbox, administered by the Nuclear Regulatory Taskforce, with authority to license Micro Modular Reactors. The sandbox would need to:</p><ul><li><p><strong>Give the licensing team who run the sandbox permission to decide which conditions are proportionate for micro reactors. </strong>There are <a href="https://www.onr.org.uk/media/gixbe2br/licence-condition-handbook.pdf">36 license conditions</a> and <a href="https://www.onr.org.uk/media/pobf24xm/saps2014.pdf">909 goals</a> in the UK&#8217;s current Nuclear Site License process, and there is no success criteria. The sandbox would allow for &#8216;technology-first&#8217; approvals; which first consider the reactor&#8217;s design basis and decide on the suitability of other principles.</p></li><li><p><strong>Align the Basic Safety Objective to the same level as background radiation in Cornwall.</strong> This would not change the developer&#8217;s responsibility to minimise radiation, but it would stop requiring paperwork once the developer has proved that <em>being next to the reactor</em> <em>if it were damaged, </em>is safer than living in Cornwall.</p></li><li><p><strong>Remove cost recovery mechanisms for the regulator. The current system has bad incentives for the regulator to extend &#8216;pre-application consultation&#8217;.</strong></p></li></ul></li><li><p>Grant planning permission and replace environmental permits for Micro Nuclear Reactors, within designated areas, provided specific environmental conditions are met, and neither the Secretary of State nor the local planning authority objects within a specified time.</p></li><li><p>Incorporate the &#8216;regulatory justification&#8217; &#8212; which currently sits within the Department for the Environment, Food, and Rural Affairs, and takes two years &#8212; into the planning decision.</p></li></ol><h3>Licensing reform</h3><p>The best way to regulate nuclear reactors is a goals-based approach. Rather than the regulator specifying how the reactor needs to be made safe (&#8220;rules-based&#8221;), the developers just have to prove that their reactor is safe. <strong>While the UK has a goals-based approach in theory, it doesn&#8217;t work like this in practice. </strong>The regulator doesn&#8217;t set out criteria for meeting these goals in advance, and has such narrowly specified success criteria, based on what they are already familiar with, that it is <em>de facto </em>rules-based. For example, at Hinkley Point C the regulator required EDF add an all-analog quadruple-backup to the control room (as in, four sets of spare equipment) despite other international nuclear regulators deeming one digital backup to be sufficient. In total, the regulator required that EDF make 7,000 design changes to a design that was already operational in France and Finland. <strong>This is </strong><em><strong>de facto </strong></em><strong>rules-based, without specifying the success criteria.</strong></p><h4><strong>Why did this happen?</strong></h4><ol><li><p><strong>The regulator is only incentivised to prevent risks from nuclear reactors, not to balance the costs and benefits of nuclear power construction.</strong></p></li></ol><p>The regulator&#8217;s <a href="https://www.onr.org.uk/">website</a> lists its mission as &#8220;to protect society by securing safe nuclear operations&#8221;. This is no expectation that it will promote, enable, or ensure the development of nuclear power. Its five statutory purposes<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> are all risk-based, meaning their goals could, in theory, be achieved without any nuclear activity at all.</p><p>There is no positive force, pulling towards nuclear getting built, to act as a counterweight. <strong>At no point is the full cost-benefit analysis happening</strong>&#8212;considering whether the benefits of additional regulation (to &#8216;safety&#8217;) outweigh the costs of it becoming prohibitively difficult to build.</p><ol start="2"><li><p><strong>The regulator has an expansive mandate and there is no oversight.</strong></p></li></ol><p>This might sound like an overstatement, but quite literally, <a href="https://www.legislation.gov.uk/ukpga/2013/32/part/3/chapter/4">Clause 78 of the Act</a> which created the regulator says that the &#8216;Principal Function&#8217; is that:</p><blockquote><p>&#8220;The ONR must do whatever it considers appropriate for the ONR&#8217;s purposes.&#8221;</p></blockquote><p>The nuclear regulator sits within the Department for Work and Pensions, so it is hardly reasonable to imagine that the Secretary of State&#8212;otherwise busy with their responsibility for <em>all benefits and the state pension</em>&#8212;would provide suitable oversight to regulator&#8217;s performance. In South Korea, where they build nuclear cheaply and quickly, the Nuclear Safety and Security Commission reports directly to the Prime Minister.</p><p>The combined effect of this incentive misalignment and expansive mandate means that companies would reasonably struggle to get the regulator to be proportionate. In <a href="https://www.onr.org.uk/publications/regulatory-reports/other-reports/onr-s-regulatory-influence-on-the-epr-design-in-the-uk/">the regulator&#8217;s response to EDF</a> publicly saying that they were required to make 7,000 design changes to Hinkley Point C, they said:</p><blockquote><p>&#8220;EDF and AREVA did not make any arguments of gross disproportion during or after the [Generic Design Assessment].&#8221;</p></blockquote><p><em>To whom </em>were EDF supposed to complain if the regulator was being grossly disproportionate? The Work and Pensions Secretary? The regulator clearly holds all the cards, and so the developer is incentivised to go along with any changes they ask for, lest it damage their chances of getting a licence.</p><ol start="3"><li><p><strong>There are no recent successful UK nuclear projects to provide a model for goals-based regulation.</strong></p></li></ol><p>The aforementioned reasons are compounded by the fact that the UK has not built a new reactor in 30 years. There are no examples of what constitutes meeting the contemporary set of goals, and the precedents from Hinkley Point C is evidently a bad guide.</p><p>This means we have to turn to case law precedent. In the UK, the law is that a marginal safety feature must be added, unless it can be deemed to be &#8216;grossly disproportionate&#8217; in costs relative to benefits. This is operationalised as when the costs to the reactor outweigh the benefits by a factor of 10 to 1.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> In practice, this means the regulator would need to simulate the view of the judge; and as is in line with their incentives, will likely default to adding more features.</p><p>By comparison, South Korea builds reactors in fleets (many at a time) which means they have very clear examples of what constitutes passing the regulations.</p><p>In aggregate, these factors lead to &#8216;goals-based&#8217; regulation morphing into a proscriptive and <strong>ever-ratching regulatory regime that prevents new construction</strong>. Consequently, in the last 35 years, the UK has decommissioned 30 reactors and added just one. While it is no individual&#8217;s fault, <strong>the nuclear regulator&#8217;s core competency has become shutting down reactors, not licensing them</strong>.</p><h3>How do we fix this?</h3><p>We need a reset; to wind back the ratchet.</p><ol><li><p><strong>Give the licensing team who run the sandbox permission to decide which conditions are proportionate for micro reactors.</strong></p></li></ol><p>There are <a href="https://www.onr.org.uk/media/gixbe2br/licence-condition-handbook.pdf">36 license conditions</a> and <a href="https://www.onr.org.uk/media/pobf24xm/saps2014.pdf">909 goals</a> that developers need to prove to get a Nuclear Site License. <em>Prima facie, </em>one might expect these will be about reactor design but they can often be organisational. For example, goal 59 is:</p><blockquote><p>&#8220;The value of safety as an integral part of good business and management practice should be reinforced through interactions between directors, managers, leaders and staff, including contractors, to establish a common purpose and collective social responsibility.&#8221;</p></blockquote><p>And license condition 3(1) is: </p><blockquote><p>The licensee shall make and implement adequate arrangements to control all property transactions affecting the site or any part of the site to ensure that the licensee remains in overall control of the site.<br><br>(Translation, if I understand correctly: prove you won&#8217;t accidentally sell the site.)</p></blockquote><p>The regulator <a href="https://www.newcivilengineer.com/latest/interview-how-the-onr-is-regulating-new-nuclear-developments-in-the-uk-28-01-2025/">has said</a>, &#8220;[W]e don&#8217;t license technologies. We license organisations to undertake a nuclear activity on a particular site.&#8221; There is some rationale to this approach, for example, it is important that the operator is capable of competently refueling their reactor. However, the organisational and site requirements would differ greatly if Radiant&#8212;who make a 1 megawatt reactor&#8212;wanted to get a license in the UK, compared to EDF building two 1650 megawatt reactors at Sizewell C.</p><p>To account for this, the licensing team should do <strong>technology-first approvals</strong>, where they first consider the design basis (as in, what the reactor is actually going to do), and then decide which of the 909 goals need to be verified based on this. At the moment, different aspects of the license application proceed in parallel, without considering whether the reactor design requires this. Taking a technology-first approach would support <strong>safety in practice </strong>rather than a box checking approach.</p><ol start="2"><li><p><strong>Align the Basic Safety Objective to the same level as background radiation in Cornwall.</strong></p></li></ol><p>The Basic Safety Objective (BSO) is the point where the regulator considers that, &#8220;beyond which further consideration of the safety case would not be a reasonable use of [the regulator&#8217;s] resources&#8221;.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> In the UK, this is set to 0.02 mSv of radiation per year.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a> This is a wholly unscientific point to choose&#8212;it is about the same as an annual roundtrip from London to New York, or eating 5.4 bananas per day for a year&#8212;it is just a quirk of the rules.</p><p>We should move the burden of proof to the <strong>Cornwall Standard.</strong><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a><strong> </strong>This would mean that&#8212;under the most unfavourable assumptions of a reactor in meltdown&#8212;the question regulators would have to answer is,<strong> &#8216;Would people receive more radiation over the course of a year, than they would receive from being in Cornwall?&#8217;</strong></p><p>Past this threshold, it should be deemed that it is a waste of public money to prove any further. This does not change the developers&#8217; responsibility to minimise radiation; it just means the regulators stop evaluating data. The Cornwall Standard would still be <em>extremely conservative</em>&#8212;by nearly a factor of 100 from my conversations with experts&#8212;on the effects on ionising radiation, but it would still be a lot of progress from the current standard.</p><ol start="3"><li><p><strong>Remove cost recovery mechanisms for the regulator</strong>.</p></li></ol><p>At present, the regulator can charge companies for &#8216;pre-application engagement&#8217; and companies cannot enter licensing until the regulator allows (known as being &#8216;license-ready&#8217;). The pre-application engagement is private, but as an example of this kind of gating; here is a quote from Rolls Royce&#8217;s <a href="https://www.onr.org.uk/generic-design-assessment/assessment-of-reactors/rolls-royce-smr/regulatory-observations-and-resolution-plans/#:~:text=Title%20Observation%20Resolution%20plan%20Closure,002%20closure%20letter">Generic Design Assessment</a>:</p><blockquote><p>&#8220;Rolls-Royce SMR Ltd should: Demonstrate the adequacy of their organisational arrangements to support the development of the E3S case for GDA. This should include roles and responsibilities, relevant processes, governance and oversight of the case.&#8221;<br><br>(Translation, if I understand correctly: prove that your company is able to write documents.)</p></blockquote><p>This might go some way to explaining why Rolls Royce&#8212;a 52 billion-pound company&#8212;still needs to receive &#163;210 million in taxpayer subsidy for its SMR. Both sides of this regulatory engagement are funded by the taxpayer, and so the cost vortex is being sustained whilst both sides think that &#8216;the other&#8217; is paying.</p><p>Instead of continuing with this broken system, we should end &#8216;cost recovery&#8217; to align the regulator&#8217;s incentives with moving at pace to license reactors, not gate access to the licensing process; and perhaps the need for subsidising the approval process will go away.</p><h3>Planning and environmental permission</h3><h4>Why is reform necessary?</h4><p>The time from initial consultation to starting construction at Sizewell C took 11 years, but for the same basic reactor in France, it took just two years.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a> During this time, there were four rounds of public consultation and the Environmental Impact Assessment <a href="https://www.gov.uk/government/publications/getting-great-britain-building-again-speeding-up-infrastructure-delivery/getting-great-britain-building-again-speeding-up-infrastructure-delivery">produced was 44,260 pages long</a>. This is clearly broken. To achieve our goal by 2027, we would need to create a new mechanism for planning, as even a Development Consent Order takes two years and still faces substantial risk of Judicial Review.</p><h4>What do we need to do?</h4><ol><li><p><strong>Give conditional planning permission to specific sites in the legislation, provided that specified environmental conditions are met, and neither the Secretary of State nor local planning authority objects. </strong>This would take inspiration from the &#8216;Renewable Acceleration Areas&#8217; created in Spain and Germany, to substitute for the Environmental Impact Assessment. We have written about this previously <a href="https://inferencemagazine.substack.com/p/getting-ai-datacentres-in-the-uk">here</a>.</p></li><li><p><strong>To ensure the new power plants have a positive environmental effect, require generous contributions to the Nature Restoration Fund. </strong>This would actually improve the efficacy of environmental mitigations: the assessment for Hinkley Point C required that EDF build an &#8216;acoustic fish deterrent&#8217; with 288 underwater speakers to prevent <a href="https://www.samdumitriu.com/p/visiting-the-worlds-most-expensive">about a trawler&#8217;s worth of fish in total</a> from being drawn into the water pumps.<br><br>This is <a href="https://www.ft.com/content/fd5e34dc-e006-491b-93b2-576e3adf45f8">&#163;100m-bat-tunnel-levels-of-ridiculous</a> and it would clearly be more efficacious to allocate this money to preserving fish populations elsewhere.</p></li><li><p><strong>Allow the local authority to keep 100% of the business rates from the new power plant to align their incentives with construction.</strong><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a></p></li><li><p><strong>Allow the developer to make payments to people who live near the reactor to compensate for the inconvenience of construction.</strong><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a></p></li></ol><h3>Regulatory Justification</h3><p>&#8216;Regulatory justification&#8217; &#8212; from the Justification of Practices Involving Ionising Radiation Regulations 2004 &#8212; requires that <em>the benefits of using ionising radiation must outweigh the costs</em>. The practical application of this in the UK, each new nuclear reactor design must show the benefits outweigh the costs, rather than saying that &#8216;nuclear power&#8217; overall must outweigh the costs. The Department for the Environment, Food, and Rural Affairs takes two years to provide a decision for each reactor, which entirely duplicates the planning process. (What is planning, if not to consider whether the benefits outweigh the costs?) Both France and Germany incorporate regulatory justification, which stems from a 1996 EU directive, into their planning process. We <a href="https://inferencemagazine.substack.com/p/getting-ai-datacentres-in-the-uk">have written</a> about the need to reform regulatory justification before, as have <a href="https://www.lexology.com/library/detail.aspx?g=8a0d1c18-822c-4013-91b4-9a0c2f6a6483">Lexology</a>, <a href="https://www.samdumitriu.com/p/how-red-tape-holds-back-nuclear-power">Britain Remade</a>, the <a href="https://institute.global/insights/climate-and-energy/revitalising-nuclear-the-uk-can-power-ai-and-lead-the-clean-energy">Tony Blair Institute</a>, and <a href="https://ukdayone.org/briefings/a-quick-win-for-green-energy-unlock-investment-for-new-nuclear">UK DayOne</a>.</p><p><strong>Regulatory justification should be incorporated into the planning verdict.</strong></p><h3>What remains?</h3><p>We have just considered what would need to be true from a planning, permitting, and licensing perspective; to enable 3 gigawatts of micro nuclear reactors by 2027. There are still other considerations&#8212;as we noted at the beginning, whether the developers are capable of delivering this; or whether the fuel supply or supply chain of skill would be able to make it in time. One concern I would not have is whether there is commercial demand for this power&#8212;I find it essentially impossible to imagine that demand for this electricity would not respond to these changes, as there is a <em>trillion dollar wave of capital expenditure for AI </em>that is principally bottlenecked by access to energy. Demand is elastic to the boldness of reforms.</p><p>So then, what is scarce? I would suggest <strong>the most scarce resource is urgency and political will.</strong> The UK was capable of getting a vaccine in under a year, and I see no reason why the same should not be true for building nuclear power by the middle of 2027. With regard to economic growth, if we had grown at 2% since 2008, and then fallen to our current level, it would be a drop tantamount to the Great Depression. The UK is <a href="https://www.ft.com/content/65b387c9-4f32-430e-877b-9985ec03f385">deindustrialising</a> because it has the highest industrial electricity prices of <a href="https://www.gov.uk/government/collections/industrial-energy-prices">any country measured by the International Energy Agency</a>, exceeding the US by a factor of 4. With regard to AI, there will be models that match human-level capabilities within the next 5 years, with an effective explosion in the cognitive workforce. And we have committed to be 95% net zero by 2030.</p><p>A good litmus test is to imagine that OpenAI wanted to build a nuclear reactor in the UK. Sam Altman has written about Greg Brockman, his cofounder, that &#8220;an average email response time of about 5 minutes to anything&#8221;, and Sam has said previously that he has written a script to see how quickly the billion dollar founders of tech companies respond to his emails versus &#8220;bad founders&#8221;; he notes, &#8220;It was a difference of minutes versus days on average response times.&#8221; The important question to ask is: <strong>does the regulator match the operating pace of companies who want to build in the UK?</strong> Currently, an industry source tells me that it can take months to get a meeting with the regulator.</p><p>To make things more concrete, people at the AI labs would  respond to an email at 10pm on a Sunday; they&#8217;d work 60 hours a week; and they&#8217;d work directly from the office of their counterparty, if they needed to, until the work was done. It seems worthwhile to consider what it would take for the state to share this level of intensity too.</p><p><strong>Slowness is a policy choice.</strong></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>The statutory purposes are: nuclear safety, nuclear site health and safety, nuclear security, nuclear safeguards, and radioactive transport safety.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p><a href="https://www.hse.gov.uk/enforce/expert/alarpcba.htm?">HSE principles for Cost Benefit Analysis (CBA) in support of ALARP decisions.</a> Note this source is not nuclear-specific, but the ALARP principle applies across industries.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p> 701, <a href="https://www.onr.org.uk/media/pobf24xm/saps2014.pdf">Safety Assessment Principles</a>, Office for Nuclear Regulation, January 2020.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>Note that this is for members of the public. For workers, the level is 0.1 mSv.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p><a href="https://ukinventory.nda.gov.uk/information-hub/about-radioactive-waste/what-is-radioactivity/#:~:text=UK%20annual%20average%20dose%20from,8">According to the Nuclear Decommissioning Authority</a>, the average background radiation in Cornwall is 7.8 mSv per year.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>While the pre-construction phase for Flammanville 3 took just 2 years, this should not imply that construction was also quick: all EPR construction has been very slow. Construction at Flammanville took 16.5 years. At Olkiluoto 3 in Finland; approval took 4.5 years, and construction took nearly 18 years. At Taishan, in China, the first EPRs were projected to take 3 years and 10 months to build, but actually took 9 years and 10 months. </p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p> Inspiration is taken from the authors of Foundations, in <a href="https://www.sambowman.co/p/one-weird-trick-to-get-data-centres">their piece on datacentres.</a> </p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>Inspiration is taken from the &#8216;Street Votes&#8217; housing policy and from Looking For Growth.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[AI and Jurisdictional Choice]]></title><description><![CDATA[Does AI do for cognitive labour what containerisation did for manufacturing?]]></description><link>https://inferencemagazine.substack.com/p/ai-and-jurisdictional-choice</link><guid isPermaLink="false">https://inferencemagazine.substack.com/p/ai-and-jurisdictional-choice</guid><dc:creator><![CDATA[Jack Wiseman]]></dc:creator><pubDate>Thu, 13 Feb 2025 22:08:21 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!9o2z!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c61df20-b545-4c7b-9acb-75940496383f_1010x1010.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Until 2022, if a company wanted to leave Delaware, they&#8217;d have to get unanimous approval from their shareholders, but following a new ruling, only a simple majority was required. Since then, Tesla and SpaceX reincorporated in Texas, Dropbox and Pershing Square Capital are reincorporating in Nevada, and Meta is also<a href="https://www.nytimes.com/2025/01/31/technology/meta-incorporation-delaware.html"> reported to be considering a move to Texas</a>. For Delaware, this is bad news: the state gets about a third of its revenue from business franchise taxes, not to mention the secondary benefits of being the place where everyone files, and disproportionately solves their corporate disputes.</p><p>Tax, too, follows this pattern: companies have relatively high levels of jurisdictional choice about where to file. For a few years, there were loopholes in Ireland, the Netherlands and some Caribbean islands, though some of these are being closed now through international treaties.</p><p><strong>Wherever firms have high levels of jurisdictional choice, there is a race to the bottom among countries and states competing for their business. Companies have relatively more power.</strong></p><p>But companies are somewhat constrained by having headquarters, management teams, shareholders, and they need to do business in major jurisdictions. They want to be on the New York Stock Exchange, even if the NYSE itself has the choice to be on a datacentre in New Jersey. In all these senses, they are strung to location.</p><p>How does AI change this?</p><p>Over time, AI agents will become an increasingly large share of economic output. While we have<a href="https://inferencemagazine.substack.com/i/155018281/ai-systems-will-be-leveraged-by-humans-mostly-not-ais-running-firms"> previously expressed some scepticism</a>, perhaps there will even be firms<a href="https://www.dwarkeshpatel.com/p/ai-firm"> entirely composed of AIs</a> in some industries. These will have much higher jurisdictional choice about where they operate&#8212;a greater fraction of &#8216;labour&#8217; can hop between datacentres, rather than being stuck in one place because of human practicalities and preferences.</p><p>AI for science is illustrative.</p><p>When AI systems can develop hypotheses, design experiments, and interpret experimental data at the level of the very best humans, then making scientific progress is no longer bottlenecked by the throughput of the most talented scientists at elite universities. The scientific process can be &#8216;deskilled&#8217;. Humans will still need to implement these experiments, as robotics aren&#8217;t good enough to fully automate the process yet.</p><p>We could quite quickly develop new tools to support the research assistants to work better. Carl Shulman has<a href="https://www.dwarkeshpatel.com/p/carl-shulman"> suggested</a>, for example, that augmented reality could abstract the requirements for <a href="https://en.wikipedia.org/wiki/Procedural_knowledge">process knowledge</a>:</p><blockquote><p>&#8220;[Y]ou could have a worker previously without training and expertise in the area who has a smartphone on a headset, and we have billions of smartphones which have eyes and ears and methods for communication for an AI to be talking to a human and directing them in their physical motions with skill as a a guide and coach that is beyond any human. They could be a lot better at telepresence and remote work and they can provide VR and augmented reality guidance to help people get better at doing the physical motions that they're providing in the construction.&#8221;</p></blockquote><p>However, even once the bottleneck of cognitive labour for science is untethered, there could be other forces keeping it tied to where it already happens: process knowledge will still be in current institutions initially, academic institutions have access to specialised equipment, and it is inconvenient to build a new lab elsewhere. So I don&#8217;t intend to make a narrow prediction about what science will look like in the near future, but rather a gesture towards the general trend: where AI systems abstract the cognitive labour from some process, or become an increasing share of output, companies will gain greater jurisdictional choice.</p><p>This is especially important from a European perspective. There&#8217;s going to be a lot of economic growth from AI &#8212; but the majority of the growth effects from general-purpose technologies come from the new products and services made possible, rather than adding it into existing processes. One has to ask, <em>why should we expect this new growth to happen in Europe? </em>When companies are going to have greater jurisdictional choice, depend less on specialised cognitive labour; decisions about where to operate will be driven comparatively by the amount of inference compute available in a market, a pro-innovation regulatory approach, and a low cost of electricity. It seems straightforward to imagine European countries finding it more difficult to compete&#8212;on regulation, energy, and abundance of inference compute. That said, <a href="https://inferencemagazine.substack.com/p/getting-ai-datacentres-in-the-uk">these things are fixable</a>.</p>]]></content:encoded></item><item><title><![CDATA[How much economic growth from AI should we expect, how soon?]]></title><description><![CDATA[Is this the steam engine, electricity, computers, or something bigger?]]></description><link>https://inferencemagazine.substack.com/p/how-much-economic-growth-from-ai</link><guid isPermaLink="false">https://inferencemagazine.substack.com/p/how-much-economic-growth-from-ai</guid><dc:creator><![CDATA[Jack Wiseman]]></dc:creator><pubDate>Fri, 17 Jan 2025 18:59:45 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc8d43a0-44c9-4a53-ae02-bb2264d08b14_1200x750.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h4>Is this the steam engine, electricity, computers, or something bigger?</h4><p>General-purpose technology revolutions have been the fundamental driver of human prosperity in the last 300 years.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a>  That these revolutions have raised the living standards of billions of people would surely indicate that, on the arrival of a new general-purpose technology, the forces for adoption must cause the world to change very quickly. But this could not be further from the truth!</p><p>The first <a href="https://en.wikipedia.org/wiki/Pearl_Street_Station">commercial power station was built in 1882</a>, and it was not until 1920&#8212;<em>nearly four decades later&#8212;</em>that electricity<em> </em>surpassed steam<em> </em>as the dominant form of horsepower in the US economy.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> In similar fashion, the microprocessor was released in 1971, but in 1990, just 20 million personal computers were sold. Among households, the pattern was consistent: reaching 50% adoption of electric lighting and a PC for the family, both took 30 years.</p><p>In the data, too, the effects are drawn out: the steam engine contributed 0.2% per year to productivity growth for 20 years, and then 0.38% per year for another 20 years in the mid-19th century.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> The largest effects on productivity from electricity took 40 years to materialise<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a>; and similarly, Robert Solow famously commented, &#8220;computers are everywhere but the productivity statistics&#8221;, which held until they finally showed up in the mid-1990s.</p><p>The important question, for our purposes, is <strong>to what extent should we expect artificial intelligence to be &#8220;just another general-purpose technology&#8221; </strong>&#8212; where growth effects very gradually over decades &#8212; <strong>or should we regard </strong><em><strong>making intelligence</strong></em><strong> as qualitatively different from previous revolutions?</strong></p><h2>Executive Summary</h2><p>The view in San Francisco is that AI will <strong>far exceed the pace and depth of change in all previous technological revolutions</strong>. This is because of a belief that AI can automate the process of invention itself. Since Bacon, the march of science depended on the actions individual inventors and small groups of researchers; but perhaps in a few years, we can create AI systems that will be capable of performing research at the level of &#8212; or indeed, much better than &#8212; the best human researchers. We can put tech progress on autopilot.</p><p>How does this arise, according to this view? First, the AI labs create an <strong>AI system capable of performing AI research</strong>, on par with their top researchers. Next, <strong>millions of instances of the &#8216;digital AI researcher&#8217; are run</strong> to make much faster research progress. These breakthroughs are applied to training the next generation of digital AI researchers, in a <strong>recursive self-improvement loop. </strong>This process leads to the creation of digital AI researchers which are <em>much </em>smarter than humans&#8212;this is &#8216;superintelligence&#8217;. In the dominant intellectual paradigm in San Francisco, this happens quickly. One important work on &#8216;takeoff speeds&#8217; towards superintelligence argued that the time between AI systems capable of performing 20% of tasks humans do, and 100% of tasks humans do, was just four years.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a></p><p>Superintelligence, as it is conceived, would have important implications for the economy: we could have an &#8216;explosion&#8217; in R&amp;D; and systems capable of performing 100% of the tasks that humans do could begin to automate the whole economy. (As the narrative goes, the superintelligence could figure out how to make robots which could perform as well as humans.) There is some academic work which investigates what happens to economic output when 100% of tasks are automatable, and many growth theory models show <strong>explosive economic growth </strong>(20% per year, or more).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a> Under some conditions there is <strong>an economic singularity</strong>, which means growth models predict infinite output in finite time.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a></p><p>On the contrary, most economists who study the impact of AI do not consider the prospect of recursive self-improvement. But most work on explosive economic growth does not deal with the microeconomic constraints of running an AI lab. This is a gap we hope to fill&#8212;providing a grounded view of what AI research automation will look like, and how this might come to affect R&amp;D and cognitive labour automation in the near future. </p><p><strong>AI research automation</strong></p><p>The most important thing to understand about AI research automation is that <strong>the AI labs are constrained by computational power to run experiments, not by researchers. </strong>A researcher from the Gemini team at DeepMind has said, &#8220;I think the Gemini program would probably be maybe five times faster with 10 times more compute or something like that&#8221;. While the cloud providers are spending enormous amounts on compute&#8212;Microsoft just announced it would spend $80 billion this year on building AI datacentres&#8212;<strong>most of this compute would be used to run inference for customers, and is unlikely to be for AI researchers to run experiments</strong>. The economics of inference for customers is very different from the economics of compute for R&amp;D: compute for experiments and training needs to be amortised across all of the inference profit margins. As we shall see, there are strong headwinds to making money selling tokens!</p><p>One of the assumptions which proponents of the Explosive Growth view often make is that a digital AI researcher will be trained on a large compute cluster, and then millions of instances will be run on the same cluster. This seems irregular to us! If the point is to recursively self-improve the AI system, but the training compute is being used for inference, where is the next generation agent going to be trained? It seems much more reasonable to imagine that ~60% of the AI labs' compute goes on serving customers, ~30% goes on training the next model, and ~10% goes on experiments. (These numbers are extremely rough guesses.) <strong>If the AI lab wants to run instances of the digital AI researcher, they will need to trade this off against experimental compute</strong>; and remember, research output is bottlenecked by experimental compute. If the digital AI researcher has equivalently good or worse ideas to the best human researcher, it makes sense to run zero copies; for it to make sense, the ideas have to be better.</p><p><strong>AI research will be automated in the future.</strong> It is reasonable to imagine that, perhaps soon, we will create a &#8216;digital AI researcher&#8217; whose research intuition&#8212;i.e. ability to predict which experiments will work&#8212;surpasses that of the best human researchers, but before then, digital AI researchers will have a bounded impact on research output, owing to the compute bottleneck. We discuss the practical challenges to increasing research output, as well as some reasons our mainline case could be wrong, in greater detail below.</p><p><strong>R&amp;D automation</strong></p><p>Concurrent to our progress on AI research automation<strong>, </strong>we want to make progress in other fields of science and technology! The opportunity is enormous&#8212;for biomedical research, clean energy, materials, synthetic biology, nanotechnology, and robotics. As with AI research, the goal is to create systems which are capable of performing all steps of the research process&#8212;generating hypotheses, designing and running experiments, and interpreting results. There are a number of challenges to scientific automation, related to the availability of data, the necessity of real-world experimentation, and so forth. It also seems reasonable to believe that academia is poorly configured to take full advantage of the opportunity which AI automation is. We expand in greater detail on both points below.</p><p>We focus on three potential fields for automation&#8212;<strong>chip research</strong> because if we are compute bottlenecked, improving our chips would help to alleviate this; <strong>robotics</strong> as improvements here could begin to automate more physical labour, and <strong>biomedical research</strong>; for effects on human wellbeing. There are different challenges in each area to automation, though in general, experimental throughput is most likely to be rate-limiting. </p><p><strong>Cognitive labour automation</strong></p><p>Thus far, chatbots and &#8216;agents&#8217; have struggled to meaningfully increase the productivity of human cognitive labour. Deploying systems is difficult right now&#8212;it requires specialised knowledge about how to build infrastructure for models. But as the models become increasingly capable of acting on long horizons, we expect most of the challenges to deployment to become diminished. We will still require people to have liability for AI systems, and in many professions, there are &#8216;embodied&#8217; complements to cognitive tasks (e.g. when a doctor has a consultation, they are both doing the diagnosis, and tailoring their explanation to the patient, and expressing care and empathy, and so on) These factors together lead us to expect that people will be managing teams of agents in their jobs&#8212;it will look like &#8216;a promotion for everyone&#8217;&#8212;rather than a lot of job losses. However, there might be some areas where production is entirely substitutable, and so jobs might be lost. To estimate the increases to output from tasks being handed off to agents, we built a growth model that shows how many tasks might be automated, how much these tasks can replace other tasks, how cheap these AI systems are, and how concentrated this is within sectors. We find that growth will be quick by historical standards, but not explosive. <strong>We expect AI will provide a 3%-9% increase to economic growth per year in the near future</strong>, and we expect it will be in the lower end of this range due to bottlenecks we discuss further in the piece. This picture will seem conservative to some&#8212;but it is worth reiterating that we will develop intelligences greater than our own, and it will radically change almost all aspects of our lives, our analysis is limited to the near-term economic picture.</p><p>There are a few variables across this whole analysis for which different assumptions would produce very different technological and economic outcomes. The most obvious is what is the inference cost of running digital researchers and cognitive labourers&#8212;if it is cheap to run both, we should expect faster research progress and we should expect greater economic growth from normal sectors of the economy. We note that it is important not to have too much confidence in a specific vision of the future; but rather see the direction of travel.</p><h2>The View From the Valley: The Economic Singularity Will Follow Superintelligence</h2><p>In the dominant intellectual framework at the AI labs, artificial intelligence is the most important technology in the history of our species.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a> The timeline looks something like this:</p><ol><li><p>The Big Bang happened (13.8 billion years ago)</p></li><li><p>Planet Earth is formed (~4.6 billion years ago)</p></li><li><p>Mammalian life began (~225 million years ago)</p></li><li><p><em>Homo sapiens </em>became the dominant species (~30,000 years ago)</p></li><li><p><em>Homo sapiens </em>build a more intelligent mind than themselves (c. 2027)</p></li><li><p>The more intelligent mind builds superintelligence (a few years later)</p></li></ol><p>It might generally be considered that <em>humans alive right now</em> are radically early&#8212;approximately <a href="https://ourworldindata.org/the-future-is-vast#our-past">108 billion humans have ever lived</a>, but if we expand to other planets, or run consciousnesses on computers, many more humans could live in the far future. For this reason, we live in <a href="https://www.cold-takes.com/most-important-century/">the most important century</a> that humans will ever live in. We live in <a href="https://michaelnotebook.com/vwh/index.html">a fragile world</a> facing many existential risks &#8212; with an estimated a 1% chance of nuclear war every year, over 200 years the chance of a nuclear war is 86.6% &#8212; and creating superintelligence, while risking existential destruction too, offers a path out of this challenge. Trillions of humans can live, on other planets or simulated on computers, and all work can be completed by robots.</p><p>We are not intending to arbitrate this diagnosis. <strong>This belief structure is much like a religion&#8212;the superintelligence has been deified, existential risk is the flood, and the AI labs are our ark.</strong></p><p>The New World will be created by an Intelligence Explosion. The anticipated narrative looks something like this:</p><ol><li><p>The human researchers will make AI agents that are capable of performing AI research.</p></li><li><p>These AI agents are run at enormous scale (millions of instances!) making much faster research progress than human researchers were.</p></li><li><p>The AI system recursively improves itself to become a &#8216;superintelligence&#8217;.</p></li><li><p>These models greatly exceed human research capabilities and are able to make other technologies, and automate all tasks in the economy.</p></li></ol><p>As a result, it is expected that all human labour (including scientific and technological progress), will be automated, and people will not need to work. We will live in &#8216;post-scarcity&#8217;&#8212;a state of complete material abundance. As an indicator of this sentiment,  Roon, a pseudonymous OpenAI researcher on X, <a href="https://x.com/tszzl/status/1852089079309176842">has tweeted</a>:</p><blockquote><p>&#8220;the future of work&#8221; there is no future of work. we are going to systematically remove the burden of the world from atlas&#8217; shoulders</p></blockquote><p>Part of this idea is that economic transformation will happen quickly&#8212;once 100% of tasks are automatable, <a href="https://www.openphilanthropy.org/research/report-on-whether-ai-could-drive-explosive-economic-growth/">this report</a> from Open Philanthropy puts a one third probability of economic growth exceeding 30% per year, and <a href="https://arxiv.org/abs/2309.11690">this paper</a> from researchers at Epoch AI say these levels of growth are &#8216;about as likely as not&#8217;. These views are based on idea-based growth models (<a href="https://en.wikipedia.org/wiki/Solow%E2%80%93Swan_model">exogenous</a>) and researcher-based growth models (<a href="https://en.wikipedia.org/wiki/Endogenous_growth_theory">semi-endogenous or endogenous</a>) which show explosive growth when AIs can substitute for humans in all economic functions.</p><p>Even without the automation of AI research, automation of a large fraction of cognitive tasks and scientific progress could lead us to explosive levels of economic growth. While lab leaders have not commented directly on economic growth, Dario Amodei (the CEO of Anthropic) has <a href="https://darioamodei.com/machines-of-loving-grace">written</a> that:</p><blockquote><p>&#8220;[M]y basic prediction is that AI-enabled biology and medicine will allow us to compress the progress that human biologists would have achieved over the next 50-100 years into 5-10 years. <strong>I&#8217;ll refer to this as the &#8216;compressed 21st century&#8217;: </strong>the idea that after powerful AI is developed, we will in a few years make all the progress in biology and medicine that we would have made in the whole 21st century&#8230;<strong>[I expect] the human economy may continue to make sense even a little past the point where we reach &#8216;a country of geniuses in a datacenter&#8217;.</strong> However, I do think in the long run AI will become so broadly effective and so cheap that this will no longer apply. At that point our current economic setup will no longer make sense, and there will be a need for a broader societal conversation about how the economy should be organized.&#8221; [emphasis ours]</p></blockquote><p>Meanwhile, Sam Altman <a href="https://moores.samaltman.com/">expressed similar sentiments</a>:</p><blockquote><p>&#8220;The technological progress we make in <strong>the next 100 years will be far larger than all we&#8217;ve made since we first controlled fire and invented the wheel</strong>.&#8230;AI will lower the cost of goods and services, because labor is the driving cost at many levels of the supply chain. If robots can build a house on land you already own from natural resources mined and refined onsite, using solar power, the cost of building that house is close to the cost to rent the robots. And if those robots are made by other robots, the cost to rent them will be much less than it was when humans made them.&#8230;Imagine a world where, for decades, everything&#8211;housing, education, food, clothing, etc.&#8211;became half as expensive every two years.&#8221; [emphasis ours]</p></blockquote><p>Do not dismiss these beliefs on the grounds they are shaped rather like a religion.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a> If nothing else, it is vitally important to understand, and take seriously, the actions of those who are building this technology. It is trivially easy to find put downs that allow one to explain away the prospect of enormous change &#8212; stories of self-importance or commercial incentive. It is much more difficult, though worthwhile, to understand the AI labs on their own terms.</p><div><hr></div><h2>Getting automation to impact growth is harder than it seems.</h2><p><em>This section is intended to provide a brief introduction to economic theory of how automation comes to increase productivity and growth. These mental models will be used throughout sections on AI research, R&amp;D, and cognitive labour.</em></p><p>In 1870, the average American worker laboured for <a href="https://www.economist.com/graphic-detail/2018/12/28/why-do-some-countries-work-longer-hours-than-others">60-70 hours per week</a>. Today, the average is <a href="https://www.economist.com/graphic-detail/2018/12/28/why-do-some-countries-work-longer-hours-than-others">35 hours</a>. We can work fewer hours to buy many more goods and services, and much better things, because workers are much more productive per hour. Tractors mean the same amount of grain can be produced by fewer farmers. And <a href="https://academic.oup.com/qje/article-abstract/70/1/65/1903777">almost all long-run growth</a> (the pie getting bigger) ultimately derives from increasing productivity.</p><p>Technology boosts productivity in two ways. First, by making tasks &#8220;cheaper&#8221; in human effort, time, or material resources; and second, by creating new tasks.</p><p>When economists talk about automation, they talk in terms of <em>tasks</em>, not jobs. Take accounting: what an accountant is doing was changed a lot, first by early computers, which could run by calculations, and then by spreadsheet software, and more recently, by &#8216;vertical SaaS&#8217; to help enterprises do bookkeeping. Sometimes this leads to a reduction in the number of people doing a job&#8212;there&#8217;s only so much accounting that a fixed number of businesses want to buy. But in other cases, the introduction of ATMs &#8212; which automates the task of giving out cash &#8212; <a href="https://www.aei.org/economics/what-atms-bank-tellers-rise-robots-and-jobs/">actually led to an increase</a> in the total number of bank tellers, as it increased the profitability of opening new branches.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-10" href="#footnote-10" target="_self">10</a> In other cases, all the tasks in a job have been completely automated, for example, lighting streetlights became unnecessary after electric street lighting was introduced.</p><p>When a task is automated, this increases productivity in two ways. First, because the tasks are cheaper, we have more resources to spend on the rest of the process, or elsewhere. Second, when one method of production gets cheaper, we tend to do more of it relative to other tasks within the same process.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-11" href="#footnote-11" target="_self">11</a> Exactly how much more we do depends on the similarity of the tasks. For example, taxis and the tube are close substitutes&#8212;if self-driving cars make getting a taxi cheaper, we should be happy because a) we&#8217;re spending less on transport, and b) because we&#8217;re using taxis in situations we otherwise wouldn&#8217;t have done because they were previously too expensive.</p><p>Where the additional resources flow to increase production depends on whether tasks are substitutes or complements. If one task is automated&#8212;say, cotton weaving&#8212;the complement to this task&#8212;for example, printing designs on cotton&#8212;becomes more valuable as a result. On the contrary, when cotton weaving was automated, the substitutes to this&#8212;handweaving&#8212;became less valuable. When a task in the production process gets automated, if the remaining tasks are very strong complements, output might not rise by much at all. For example, if there is a packaging machine for cotton goods which is already operating at its limit, the automation of weaving will save resources, but cannot increase the output of cotton goods.</p><p>The same pattern applies at the level of the economy too! If there is more extensive automation in some sectors, the price of goods produced in that sector will fall, and so that sectors&#8217; share of GDP (total output) grows less. This means that GDP ends up being composed of things which are essential, and yet hard to automate. Agriculture used to be <a href="https://www.nber.org/system/files/chapters/c8007/c8007.pdf">90% of GDP</a>, but since mechanisation, it has shrunk as a fraction of GDP, to just <a href="https://www.ers.usda.gov/data-products/chart-gallery/gallery/chart-detail?chartId=58270#:~:text=According%20to%20data%20from%20the,0.8%20percent%20of%20U.S.%20GDP.">0.8%</a>. Total output is bottlenecked by that which is essential &#8212; <a href="https://www.aei.org/carpe-diem/chart-of-the-day-or-century-8/">healthcare, education, housing</a> &#8212; but hard to automate! This is known as the <a href="https://en.wikipedia.org/wiki/Baumol_effect">Baumol effect</a>.</p><p>The important things to keep in mind when considering any automation are: how much does this automation directly reduce costs, and to what extent is output bottlenecked by this, or another factor?</p><div><hr></div><h2>AI research will be automatable, but the practical details will matter a lot.</h2><p>The stated goal of much AI research is to make an AI researcher. The hope is that by automating the work of human AI researchers, we can make faster progress in AI research. This is for two reasons:</p><ol><li><p>Because we can run more copies of the AI researcher than we can have human researchers.</p></li><li><p>Because we can engineer the digital AI researcher to continue getting cleverer than the human researcher, in an essentially unbounded way.</p></li></ol><h3>What does progress towards the AI scientist look like?</h3><h4>To get the digital AI scientist, the system must be able to perform all the sub-tasks involved in AI research.</h4><p><strong>What does an AI researcher do?</strong></p><p>From a series of interviews with AI researchers, Epoch AI <a href="https://epoch.ai/files/Interviewing_AI_researchers_on_automation_of_AI_R_D.pdf">created a taxonomy</a> of the tasks involved in AI research. In the simplest model, AI researchers create hypotheses, design experiments, run the experiments, analyse the results, and repeat this cycle.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!02MH!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa0b30dd-8633-444d-b953-2f5d95b0c36a_1600x1472.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!02MH!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa0b30dd-8633-444d-b953-2f5d95b0c36a_1600x1472.png 424w, https://substackcdn.com/image/fetch/$s_!02MH!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa0b30dd-8633-444d-b953-2f5d95b0c36a_1600x1472.png 848w, https://substackcdn.com/image/fetch/$s_!02MH!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa0b30dd-8633-444d-b953-2f5d95b0c36a_1600x1472.png 1272w, https://substackcdn.com/image/fetch/$s_!02MH!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa0b30dd-8633-444d-b953-2f5d95b0c36a_1600x1472.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!02MH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa0b30dd-8633-444d-b953-2f5d95b0c36a_1600x1472.png" width="1456" height="1340" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/aa0b30dd-8633-444d-b953-2f5d95b0c36a_1600x1472.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1340,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!02MH!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa0b30dd-8633-444d-b953-2f5d95b0c36a_1600x1472.png 424w, https://substackcdn.com/image/fetch/$s_!02MH!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa0b30dd-8633-444d-b953-2f5d95b0c36a_1600x1472.png 848w, https://substackcdn.com/image/fetch/$s_!02MH!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa0b30dd-8633-444d-b953-2f5d95b0c36a_1600x1472.png 1272w, https://substackcdn.com/image/fetch/$s_!02MH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa0b30dd-8633-444d-b953-2f5d95b0c36a_1600x1472.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-12" href="#footnote-12" target="_self">12</a></p><p>There are many valuable questions which could provide a clearer picture of what it would take, and what it means, to automate this:</p><ul><li><p>How much time do AI researchers spend between hypothesis generation, designing experiments, and analysing results?</p></li><li><p>When AI researchers reflect on their own cognition while generating hypotheses, what kinds of reasoning are they doing?</p></li><li><p>How much time do they spend waiting for the results of experiments?</p></li><li><p>(I could go on&#8230;)</p></li></ul><p>There are very few public resources which deal with automating AI research at frontier labs, but more specific materials would make predictions of transformative change much easier. For most of this analysis, we rest on <a href="https://www.dwarkeshpatel.com/p/sholto-douglas-trenton-bricken">an enormously valuable interview</a> on the Dwarkesh Podcast with Sholto Douglas (a Google DeepMind researcher) and Trenton Bricken (an Anthropic researcher).</p><h4>Current systems are getting better at ML engineering, but performance struggles over longer horizons&#8230;</h4><p><strong>How do state-of-the-art models perform on our current tests of AI R&amp;D?</strong></p><p>We develop tests of AI research, or benchmarks, which can give us a smooth function of how much progress we are making towards the capability. Good benchmarks give you a score out of 100 on a diverse range of tests that most closely mirrors the capability in the real world. There are three main benchmarks which test a model&#8217;s AI Research abilities<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-13" href="#footnote-13" target="_self">13</a>: OpenAI&#8217;s <a href="https://openai.com/index/mle-bench/">MLE-bench</a>; METR&#8217;s (a non-profit model evaluator) <a href="https://arxiv.org/abs/2411.15114">RE-bench</a>; and OpenAI&#8217;s <a href="https://www.swebench.com/">SWE-bench Verified</a>.</p><p><strong>MLE-bench </strong>tests the models against 75 ML engineering questions from online competitions (Kaggle). The latest public scores on this benchmark are for o1-preview, which lags o1 and o3. O1-preview performed in the top 40% of humans who had completed these ML engineering tasks on 16.9% of occasions. We should expect o3 to perform <strong>significantly </strong>better on this benchmark.</p><p><strong>RE-bench </strong>tests the models against seven ML engineering tasks relevant for frontier R&amp;D.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-14" href="#footnote-14" target="_self">14</a> In the evaluation, o&#8211;1 preview and Claude 3.5 Sonnet outperformed human experts with a two-hour time budget, but model performance asymptotes while human performance continues to rise with additional hours.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!RV7f!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc8d43a0-44c9-4a53-ae02-bb2264d08b14_1200x750.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!RV7f!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc8d43a0-44c9-4a53-ae02-bb2264d08b14_1200x750.png 424w, https://substackcdn.com/image/fetch/$s_!RV7f!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc8d43a0-44c9-4a53-ae02-bb2264d08b14_1200x750.png 848w, https://substackcdn.com/image/fetch/$s_!RV7f!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc8d43a0-44c9-4a53-ae02-bb2264d08b14_1200x750.png 1272w, https://substackcdn.com/image/fetch/$s_!RV7f!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc8d43a0-44c9-4a53-ae02-bb2264d08b14_1200x750.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!RV7f!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc8d43a0-44c9-4a53-ae02-bb2264d08b14_1200x750.png" width="1200" height="750" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cc8d43a0-44c9-4a53-ae02-bb2264d08b14_1200x750.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:750,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!RV7f!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc8d43a0-44c9-4a53-ae02-bb2264d08b14_1200x750.png 424w, https://substackcdn.com/image/fetch/$s_!RV7f!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc8d43a0-44c9-4a53-ae02-bb2264d08b14_1200x750.png 848w, https://substackcdn.com/image/fetch/$s_!RV7f!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc8d43a0-44c9-4a53-ae02-bb2264d08b14_1200x750.png 1272w, https://substackcdn.com/image/fetch/$s_!RV7f!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc8d43a0-44c9-4a53-ae02-bb2264d08b14_1200x750.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-15" href="#footnote-15" target="_self">15</a></p><p>It is noteworthy that, on the task of optimising a kernel for latency, one of the models was able to find a solution with lower latency than the best solution from any human researcher in the benchmarking.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-16" href="#footnote-16" target="_self">16</a> As with MLE-bench, we expect that o3 performs significantly better than o1-preview. To provide an indicator to the trajectory of progress, Sholto Douglas <a href="https://x.com/_sholtodouglas/status/1860611129825022169">tweeted</a> that within six months, they expect state-of-the-art models to outperform human researchers with a time-budget of four hours.</p><p><strong>SWE-bench verified </strong>tests model performance on real-world software engineering tasks. These do not reflect the models&#8217; ability to do ML research, but provide a more general indicator of their coding ability. OpenAI&#8217;s o3 release in December 2024 took state-of-the-art performance to <a href="https://venturebeat.com/ai/openai-confirms-new-frontier-models-o3-and-o3-mini/">71.7%</a>. We predict this benchmark will be saturated (i.e. so close to 100% that it no longer useful distinguishes between models) in 3 to 6 months.</p><p>Finally, it is notable that <strong>ARC-AGI-PUB</strong>&#8212;a benchmark of models&#8217; visual reasoning abilities&#8212;has become saturated. It measures performance on a series of visual reasoning puzzles (a bit like non-verbal reasoning tests, for those who went to school in the UK). The AI models have historically struggled with these problems, but humans would find them trivially easy to solve. GPT-4o, released in May 2024, scored just 5% on the benchmark; o1-preview scored 13.3%, but now, o3 is able to score <a href="https://venturebeat.com/ai/openai-confirms-new-frontier-models-o3-and-o3-mini/">88%</a>.</p><p>At this point, it is useful to reflect on what it would mean for all these benchmarks to become saturated. These are necessarily imperfect snapshots of <em>what it is to do AI research</em>. These tasks do not map comprehensively to Epoch AI&#8217;s taxonomy above. As <a href="https://x.com/BethMayBarnes/status/1860065450824204686">one of the creators of RE-bench notes</a>, their benchmark does not capture the models&#8217; ability to interact with large and messy codebases, and make compute allocation decisions. However, these benchmarks provide a guide for the rate at which models are becoming better at ML engineering.</p><p>Improving these capabilities technically will depend on training systems for better long-horizon task performance. For more detailed coverage, see these pieces in Inference.</p><h4>Partial automation is unlikely to provide much productivity uplift to AI research.</h4><p>As we have <a href="https://docs.google.com/document/d/1mRzI1gYf_0Kmag23JMDrOlTCAl6r7TK4JynZZDzztXw/edit?tab=t.0#heading=h.fovupzb7qauf">mentioned earlier</a>, the relevant question to consider for automation is: how much does automating this step actually impact output?</p><p>In this case, we want to know how much research output is increased by an agent (&#8220;the proto-AI researcher&#8221;) which is:</p><ul><li><p>Less good than human researchers at generating hypotheses;</p></li><li><p>But can do software or machine learning engineering faster than humans, at or above the level of human research engineers.</p></li><li><p>And it can create visualisations of research results with preliminary analysis, to present to human researchers.</p></li></ul><p>We do not expect this &#8220;proto-AI researcher&#8221; to increase output much for the following reasons.</p><p><strong>We strongly expect that the output of AI research labs is bottlenecked by compute. </strong>In the Dwarkesh interview, Sholto Douglas <a href="https://www.dwarkeshpatel.com/p/sholto-douglas-trenton-bricken">has said</a>, &#8220;I think the Gemini program would probably be maybe five times faster with 10 times more compute or something like that.&#8221; It is notable also that, according to Situational Awareness (as of May 2024), &#8220;GDM is rumoured to have way more experimental compute than OpenAI&#8221;. Perhaps the returns to marginal experimental compute are even more dramatic at other AI labs.</p><p>But there is a more simple &#8216;outside view&#8217; argument, for why we should expect AI research labs to be compute bottlenecked. <strong>Put yourself in the shoes of a Chief Scientist&#8212;if you aren&#8217;t saturating your experimental compute, you should be trying extremely hard to!</strong></p><p>Maintaining experimental compute clusters cost the AI labs billions of dollars. In comparison, AI researcher salaries are just a few hundred thousand, to a few million dollars a year. If your researchers don&#8217;t have enough ideas to saturate the compute you have, you should hire more researchers! If your best researchers have too many ideas without the time to implement them, you should hire more research engineers to help them do this! Being in a regime where you <em>aren&#8217;t constrained by compute</em> means the bottleneck is something else, which are worse problems to have.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-17" href="#footnote-17" target="_self">17</a></p><p><strong>Using the proto-AI researcher for implementing experiments would require researchers to change their workflows, which could limit, or even reduce, their research output. </strong>Researchers have spent their entire working lives honing their research process, and experimental design is hard. If their AI lab asked their researchers to switch to a new workflow for implementing experiments where they have to pre-specify their view of how to do an experiment, in natural language, this could dampen their creativity, or they could spend time correcting models which did not implement the experiment as well as they would have. For some researchers, <em>writing the experiment is thinking about the experiment</em>. Breaking up this flow could limit, or even negate, the benefits of faster implementation.</p><p>However, this view of limited progress could be wrong for a few reasons.</p><p><strong>The proto-AI researcher could partially automate benchmark creation. </strong>It might sound weird to outsiders, but one of the places where AI labs are most bottlenecked right now is after an experiment, how do they work out if the change they made<em> actually improved the model? </em>They use benchmarks, like the ones we&#8217;ve already discussed, but finding good tests has been getting increasingly hard. Zhengdong Wang, a Google DeepMind researcher, has an excellent section in his excellent <a href="https://zhengdongwang.com/2024/12/29/2024-letter.html">end of year letter</a>, on the problem of working on a poorly specified goal like &#8216;make this model generally intelligent&#8217;:</p><blockquote><p>But how does [the researcher] know which experiment is better? In the past, evaluation was easy because the desired result was clear. If one model won more games of chess, or predicted a protein structure with higher accuracy, then it was better. Today, &#8220;better&#8221; is vaguer and slower to get than ever before. Our researcher can interact with a model for a long time, or look at which model users like more when he deploys it. But to do effective research, he needs fast (read: automated) evaluations. So he resorts to a test or benchmark (colloquially, an &#8220;eval&#8221;) that is unambiguous enough, fast enough, and a good enough approximation of what he means by &#8220;better.&#8221; Concretely, sipping his coffee, our researcher is looking at a plot where training progress is on the x-axis, and performance on a test is on the y-axis. He wants performance to go up as training progresses.</p><p>&#8230;<br><br>In fact, you might even say that the only time AI researchers are doing AI research is when they choose the evaluation. The rest of the time, they&#8217;re just optimizing a number.</p></blockquote><p>This matches with a lot of what we&#8217;ve heard from people at the labs&#8212;the multiple choice questions it is possible to generate are being saturated, and from here, we will need benchmarks of longer-horizon tasks. These benchmarks will need to provide agents with an environment to act in, a well-defined task that provides a signal of their intelligence, with a smooth function that summarises how well they are performing. (To think about how difficult this is, work backwards: what test can you design to show &#8220;this model is 50% &#8216;good at science&#8217;&#8221;?) From conversations with people who make benchmarks, they expect it could be possible to automate quite a lot of their work&#8212;setting up environments, designing verification tasks for the models, and orchestrating agents at scale. However, they also stressed there are compute constraints on running large-scale long-horizon benchmark tasks, and that models will need <em>even harder </em>tests, automating which might jump ahead of current model capabilities.</p><p>In short, if the researchers were able to get clearer and deeper signal about how their models are improving in the domains they care about, as quickly as possible, it could well accelerate iteration speed on the incremental improvements to the models.</p><p><strong>The human researchers could be &#8216;freed up&#8217; to spend more time on other tasks</strong>, like thinking about better experiments to run, or reading more literature which could sow the seeds for better ideas in the future, or think more deeply about their experimental results and what might be happening inside the models.</p><p>However, we are skeptical, because we suspect that actually writing experiments is a small fraction of the job. <a href="https://www.dwarkeshpatel.com/p/sholto-douglas-trenton-bricken">Sholto notes</a>,</p><blockquote><p>&#8220;People have long lists of ideas that they want to try. Not every idea that you think should work, will work. Trying to understand why that is is quite difficult and working out what exactly you need to do to interrogate it. So a lot of it is introspection about what's going on. It's not pumping out thousands and thousands and thousands of lines of code.&#8221;</p></blockquote><p>If implementing ideas for experiments is only a small fraction of time to begin with, speeding this process up doesn&#8217;t create much additional time for more thinking about experiments to run.</p><h4>Complete automation could be bottlenecked by ideation and research taste.</h4><p>In models of automation, there are a small number of tasks which are the last to be automated. These are known as &#8216;holdout tasks&#8217;. Whether there will be holdout tasks, what these will be, and how long they might <em>hold out </em>for, are important for understanding the output of &#8216;proto-AI researchers&#8217;.</p><p>Longer explanations of technical progress have been covered here and here in Inference, but in the briefest manner: the systems which AI labs are training will be trained to perform long-horizon tasks. This requires improving the models&#8217; ability to maintain goal-directedness and coherence (and not to drift off track, as is sometimes observed in weaker models). This also requires error detection and recovery, as weaker models typically get stuck in loops, making the same mistake. As part of this regime, the models are being trained to think for longer, in order to improve their reasoning and planning capabilities.</p><p>We are currently in an &#8216;inference-time compute&#8217; overhang, which means we have the capacity to increase the amount of compute which AI systems are using during inference, for greater capabilities. The relevant question, for our purposes on the complete automation of AI research; is where does the overhang end?</p><p>It could be the case that we have all the relevant components of creating an AI researcher <em>within this current overhang</em>. Perhaps all that it takes to make AI researchers with better ideas than human researchers is to scale up the models&#8217; ability to think for a long time, and give them good examples of human researchers research ideas for a given set of evidence. On the other hand, there could be some cognitive tasks which the current overhang is unable to capture, and so output remains bottlenecked by these. For example, perhaps the lead researchers who set the research direction of the lab, and have to set plans for an extended period are engaging in a type of reasoning which is inaccessible; or perhaps the digital AI researchers are unable to reach the reliability at generating good ideas of human researchers.</p><p>For the view that it is possible within the paradigm; see &#8216;AGI is an engineering problem&#8217;.</p><h4>It is unlikely there will be a discreet moment when we &#8216;have&#8217; the AI researcher. We expect it to emerge over time.</h4><p>When the models&#8217; ideas for experiments are 90%-as-good-as the best human researchers&#8217;, they will be used 0% of the time. But once the models&#8217; ideas can sometimes generate ideas 105%-as-good-as the best human researchers&#8217;, the human researcher should notice and implement them. Despite this, knowing in practice when the model&#8217;s ideas are better than a researchers&#8217; seems to be particularly difficult. AI labs could not gamble on automating AI research prematurely, only to discover their agents&#8217; ideas are worse than human researchers at a competing AI lab. Discovering the models are better at thinking of ideas is likely to be a gradual process&#8212;when their ideas are roughly as good as the best researchers', deciding which side of the distribution around 100% they fall will be very difficult for the researchers.</p><h3>To what extent can AI labs maintain an experimental compute budget?</h3><p>Experimental compute is central to our narrative. If it can have such dramatic effects on research output &#8212; that 10 times more compute means 5 times more progress &#8212; then sustaining as much compute as possible is vital for all research labs. Whoever has the most compute might even be the decisive factor, for who reaches the AI researcher first.</p><p>In the headline there are enormous $ figures for big tech companies spending on AI infrastructure &#8212; just last week, Microsoft announced that it would spend $80 billion on building new datacentres in 2025 &#8212; but most of this will be to run inference for customers through their product suite or cloud, and not for experimental or training compute for AGI labs.</p><h4>The economics of compute for R&amp;D are different from the economics of serving models to customers.</h4><p>When GPUs are used for serving customers, the goal is to generate as much surplus as possible. Because GPUs are very expensive &#8212; electricity is <a href="https://epoch.ai/blog/can-ai-scaling-continue-through-2030#:~:text=We%20investigate%20the%20scalability%20of,likely%20be%20feasible%20by%202030.">only ~10-15%</a> of the total cost of ownership &#8212; you do not want them to be idle. If you want to have the biggest surplus, you'll need to run a) as many GPUs as possible, at b) as high utilisation as possible, whilst c) providing all customers with a suitable level of interactivity. This is extremely difficult &#8212; how do you split the model across multiple GPUs to tradeoff throughput and interactivity? How do you forecast the number of GPUs you'll want in 4 years' time (the horizon for making AI infrastructure decisions)? How do you know <em><a href="https://www.reddit.com/r/mlscaling/comments/1eyophn/hardware_hedging_against_scaling_regime_shifts/">the kind of hardware</a></em> you will need to run the models of 2028? To what extent will we make inference efficiency gains, so that demand for inference can be satisfied with a much smaller number of GPUs than it would take today?</p><p>On the other hand, R&amp;D compute is about <em>spending the surplus</em>. The goal of R&amp;D is to create new models, which will maximise your future surplus:</p><ul><li><p>Either because the models are more widely useful, and so you can sell more tokens;</p></li><li><p>Or because they are differentially capable, so you can charge more relative to the cost of inference;</p></li><li><p>Or because they are more efficient to run for a given capability level, so you take more home as surplus.</p></li></ul><p><strong>If R&amp;D compute is about spending the surplus you&#8217;ve generated, then your total R&amp;D compute needs to be amortised over all your inference.</strong></p><p>The total amount which needs to be amortised is rising over time. The <a href="https://ifp.org/future-of-ai-compute/">table below</a> from the <em>Institute for Progress</em> shows the growth in training compute over time. The computational power dedicated to the largest training runs will be 100 times as large in 2030, as in 2026. (Note that while pre-training scaling laws might well be slowing, more compute can be applied during post training. <a href="https://semianalysis.com/2024/12/11/scaling-laws-o1-pro-architecture-reasoning-training-infrastructure-orion-and-claude-3-5-opus-failures/">SemiAnalysis predicts</a> post-training FLOP will exceed pre-training FLOP in future.)</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!H2n3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F450499b8-b304-4a8b-8d07-5cf13fad6c16_1432x558.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!H2n3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F450499b8-b304-4a8b-8d07-5cf13fad6c16_1432x558.png 424w, https://substackcdn.com/image/fetch/$s_!H2n3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F450499b8-b304-4a8b-8d07-5cf13fad6c16_1432x558.png 848w, https://substackcdn.com/image/fetch/$s_!H2n3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F450499b8-b304-4a8b-8d07-5cf13fad6c16_1432x558.png 1272w, https://substackcdn.com/image/fetch/$s_!H2n3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F450499b8-b304-4a8b-8d07-5cf13fad6c16_1432x558.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!H2n3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F450499b8-b304-4a8b-8d07-5cf13fad6c16_1432x558.png" width="1432" height="558" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/450499b8-b304-4a8b-8d07-5cf13fad6c16_1432x558.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:558,&quot;width&quot;:1432,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!H2n3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F450499b8-b304-4a8b-8d07-5cf13fad6c16_1432x558.png 424w, https://substackcdn.com/image/fetch/$s_!H2n3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F450499b8-b304-4a8b-8d07-5cf13fad6c16_1432x558.png 848w, https://substackcdn.com/image/fetch/$s_!H2n3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F450499b8-b304-4a8b-8d07-5cf13fad6c16_1432x558.png 1272w, https://substackcdn.com/image/fetch/$s_!H2n3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F450499b8-b304-4a8b-8d07-5cf13fad6c16_1432x558.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-18" href="#footnote-18" target="_self">18</a></p><p>The total cost of ownership for an H200 is roughly $10.5k per month.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-19" href="#footnote-19" target="_self">19</a> (edit: correction, this previously said H100, but should have said H200; the TCO of an H100 is ~$9k/p.m.) For a 100k cluster, the annual cost will be roughly $1.6 billion. For the median 2028 cluster, it will be roughly $8.9 billion annually for ownership (note this is not capex!). On top of this, add experimental compute. All algorithmic improvements need to be tried at multiple increments of scale, and so the training compute will need to be at least close to the largest cluster. We estimate, with low confidence, that all experimental compute, as well as for evals and for safety research might be the same as the training cluster. And so R&amp;D compute in 2028 might be a $15 billion to $20 billion expense for the AI labs.</p><p>Paying for this means selling some tokens!</p><p>An important question to consider: what is the economically-useful life of a model? We will argue that&#8230;</p><h4>The economically-useful life of a model is short.</h4><p><strong>&#8216;Frontier&#8217; capabilities seem to get commoditised quickly, which hurts margins. </strong>Thus far, OpenAI have generally released the most powerful capabilities first. But not long after, other AI labs have released similarly powerful models.</p><p>GPT-4 was <a href="https://arxiv.org/abs/2303.08774">released in March 2023</a>, although it <a href="https://openai.com/product/gpt-4">finished training</a> in August 2022, and just four months after its release, Anthropic released <a href="https://www.anthropic.com/news/claude-2">Claude 2</a> and Meta released <a href="https://arxiv.org/abs/2307.09288">Llama 2</a>. <em>To what extent</em> the capabilities of GPT-4 were commoditised at this point is debatable: Llama 3 had a <a href="https://arxiv.org/abs/2407.21783">markedly worse</a> HumanEval score (a test of coding ability), but the point is that directionally, in just 4 months, the competitive differentiation of GPT-4 was diminished.</p><p>GPT-4 Turbo was released in November 2023, and by March 2024, Anthropic released Claude 3, xAI released Grok-1, and then in April, Meta released Llama 3. All of these models had roughly equivalent benchmark scores.</p><p>Finally, OpenAI released GPT-4o in May 2024; Anthropic followed in June with Claude 3.5 Sonnet, and Meta followed in July with Llama 3.1.</p><p>When the leading model is clearly differentiated, the AI lab who made it will be able to make excess profits; but when these capabilities are commoditised, their margin is competed away. The less margin there is, the more difficult it is to amortise the cost of training new models (and the more one depends on the size of one's customer base).</p><p>However<strong>, commoditisation of capabilities could end</strong>. In the previous paradigm, when there was only a single axis for improvement (base model scale) there was natural convergence towards similar levels of capabilities. In the new paradigm of inference-time scaling, the 'returns to ideas' rise&#8212;first, everyone needs to make the leap to follow OpenAI in being able to scalably apply more compute at inference time, and second, there are many different types of RL which could have this affect. If the labs are decorrelated in their approaches, it is plausible to imagine that their models could have more heterogeneous capabilities. Additionally, more advanced post-training techniques offer the potential for more advanced &#8216;personality&#8217; elicitation from the systems: people generally seem to prefer Claude 3.5 Sonnet&#8217;s style, tone of voice, and writing ability over other models. On the other hand, even if the techniques labs choose are different, they will want to train their models towards the same tasks&#8212;being good at coding, being able to complete tasks on a computer&#8212;and so even with different research approaches, they can end up in the same place.</p><h5><strong>Irrespective of whether the software layer commoditises, hardware capabilities will take much longer to compete away.</strong></h5><p>Once Google DeepMind is able to unlock Chain of Thought models, we expect they will have strong cost advantages for running large models on TPUs, against others running on GPUs. The TPUv6&#8217;s operate in a pod of 256 other chips, while NVIDIA Hoppers, B100, and B200 are only able to maintain a pod of 8 chips. Even the next-generation GB200 only has a pod size of 72. Larger pods make more parallelism schemes possible (i.e. you can split the model across a wider number of TPUs, with more creative configurations).</p><p>(As a technical detail: CoT costs do not scale linearly, so as sequence lengths at inference get longer, the problem gets worse.)</p><p><strong>Switching models is easy, which means margins are more competitive. </strong>Moving between model providers is as simple as editing a line of code to change the API call. From the conversations we&#8217;ve had, if your scaffolding is built correctly, changing the base model does not cause this to break. Perhaps this changes, as model providers build developer tools and add parts of the bundle to keep you in their system, but it is at least not true for now.</p><p><strong>At present, there is no market for &#8216;non-frontier&#8217; models. </strong>AI systems do not seem to have reached the efficient frontier of latency, cost or performance&#8212;there&#8217;s so much further to go.<strong> </strong>It also is worth mentioning that typically, how the smaller models with lower latency and cost are created is distilling the larger model (a la o1-mini is to o1) rather than having a different training process, or 'falling off' the frontier.</p><p><strong>Inference efficiency gains mean that prices per token fall, so R&amp;D compute has to be amortised over a larger number of tokens.</strong></p><p>Over the past two years, token prices for GPT-4 series models have collapsed by roughly 240x, per the <a href="https://x.com/eladgil/status/1827521805755806107">chart from Elad Gil</a> below. What is driving this? We would guess that it is not <em>entirely inference optimisations</em>. Perhaps in the early months of GPT-4, OpenAI had limited compute resources for serving models, so the high token prices were a form of rationing. However, this provides some directional indicator for the kinds of inference gains made over the period. Perhaps it is an extremely obvious point, but if token prices fall by 240 times, whilst margins hold constant, then making the same amount of revenue at the end of the period would require selling 240 tokens for every token you sold at the start of the period.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!0jWm!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8ab170d-f082-4f73-8f91-e8f580f184fa_1180x804.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!0jWm!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8ab170d-f082-4f73-8f91-e8f580f184fa_1180x804.png 424w, https://substackcdn.com/image/fetch/$s_!0jWm!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8ab170d-f082-4f73-8f91-e8f580f184fa_1180x804.png 848w, https://substackcdn.com/image/fetch/$s_!0jWm!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8ab170d-f082-4f73-8f91-e8f580f184fa_1180x804.png 1272w, https://substackcdn.com/image/fetch/$s_!0jWm!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8ab170d-f082-4f73-8f91-e8f580f184fa_1180x804.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!0jWm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8ab170d-f082-4f73-8f91-e8f580f184fa_1180x804.png" width="1180" height="804" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d8ab170d-f082-4f73-8f91-e8f580f184fa_1180x804.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:804,&quot;width&quot;:1180,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:272805,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!0jWm!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8ab170d-f082-4f73-8f91-e8f580f184fa_1180x804.png 424w, https://substackcdn.com/image/fetch/$s_!0jWm!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8ab170d-f082-4f73-8f91-e8f580f184fa_1180x804.png 848w, https://substackcdn.com/image/fetch/$s_!0jWm!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8ab170d-f082-4f73-8f91-e8f580f184fa_1180x804.png 1272w, https://substackcdn.com/image/fetch/$s_!0jWm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8ab170d-f082-4f73-8f91-e8f580f184fa_1180x804.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-20" href="#footnote-20" target="_self">20</a></p><p>To step back from the analysis of inference costs, what these parameters have established in aggregate, is that each model has a window of opportunity to make a surplus that can be used to pay for R&amp;D for future models. These windows seem to be short, with limited opportunity for margin throughout, and amortisation getting more difficult over time. Superintelligence seems like a bad business.</p><h4>The labs will need to make products, not sell tokens, to fund R&amp;D.</h4><p>In order for this to work, AI labs need to get out of the token business, and start selling automated tasks or agents which can be priced in terms of their labour-equivalent.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-21" href="#footnote-21" target="_self">21</a> More on the impact of <a href="https://docs.google.com/document/d/1mRzI1gYf_0Kmag23JMDrOlTCAl6r7TK4JynZZDzztXw/edit?tab=t.0#heading=h.s8erklwd3iif">AI systems on cognitive labour in a later section</a>.</p><h3>Allocating the experimental compute budget is difficult.</h3><p>Most visions of fast AI progress, in our opinion, assume away the problem of compute budgets. In practice, experimental compute is hard-won; and to what extent computational power can be dedicated towards particular goals determines a great deal about how quickly they are achieved.</p><h4>The amount of experimental compute that AI labs have places limits on the size of their human AI researchers teams.</h4><p>Even without the digital AI researcher, if an AI lab was going to hire the marginal human research, they would need to divide compute between &#8216;n + 1&#8217; researchers. This means that some fraction of the compute from your best researchers will be reassigned to the new researcher. Because there are extremely high returns to research talent, <em>and compounding benefits to intuition and research taste from having run lots of experiments</em>; it is quite likely that it does not make sense to add more human researchers at all. In fact, labs might want to centralise lots of compute behind a very small number of very talented research scientists, and have the great majority of people improving their efficiency; a bit like a Megazord from Power Rangers. For an AI researcher to get involved in the research pipeline, their ideas have to surpass what the best researchers could otherwise spend the compute on.</p><h4>Instances of the digital AI researcher will not be run on the cluster they were trained on.</h4><p>The arguments for the Intelligence Explosion are premised on the idea that we will run the instances of the AI researcher <em>on the cluster they were trained on</em>. This does not seem to match how we&#8217;d imagine AI lab compute budgets work, nor how it would be optimal for them to work. For starters, where would the next model be trained? We would expect that the compute budget of an AI lab looks something like this: ~60% of compute on serving customers&#8217; models; ~20% on training compute for the biggest run (~10% on pre-training, ~10% on post-training), ~10% on experimental compute, and ~10% on &#8216;other things&#8217; (e.g. synthetic data generation, safety research, evaluations). Therefore, if the labs wants to run instances of the AI scientist, these will need to be traded off against experimental compute. The question then becomes: what is the marginal use of this compute? The next-best experiments of the best human researchers, or instances of the AI scientist to <em>think about what experiments to run?</em></p><p>In this view, the digital AI researcher's ideas would need to generate much better ideas than those of the human researchers, otherwise it would not make sense to use up the compute.</p><h4>The early inference costs of the AI scientist are likely to be very high, though we should expect them to fall dramatically.</h4><p>The narratives for the Intelligence Explosion rely upon inference costs for the AI researcher being trivially low. Recent examples of Chain of Thought models have shown them using enormous amounts of inference compute&#8212;if it cost <a href="https://news.ycombinator.com/item?id=42473321">$3400</a> on average to solve each ARC-AGI benchmark prize, and we use the token prices of o1-preview, it took roughly 128,000 tokens to solve problems which are trivially easy for humans to solve.</p><p>A quick technical detail: recall that inference costs of CoT models do not scale linearly, as the model is forced to parallelise across multiple pods.</p><p>If we are to get additional capabilities through scaled Chain of Thought, this description of the AI researchers in <a href="https://situational-awareness.ai/from-agi-to-superintelligence/">Situational Awareness</a> would seem to have astronomical inference costs&#8230;</p><blockquote><p>&#8220;[T]hey&#8217;ll [each of the 100 million AI researcher] be able to get incredible ML intuition (having internalized the whole ML literature and every previous experiment every run!) and centuries-equivalent of thinking-time to figure out exactly the right experiment to run, configure it optimally, and get the maximum value of information; they&#8217;ll be able to spend centuries-equivalent of engineer-time before running even tiny experiments to avoid bugs and get them right on the first try; they can make tradeoffs to economize on compute by focusing on the biggest wins; and they&#8217;ll be able to try tons of smaller-scale experiments (and given effective compute scaleups by then, &#8220;smaller-scale&#8221; means being able to train 100,000 GPT-4-level models in a year to try architecture breakthroughs).&#8221;</p></blockquote><p>It also seems to endow the AI researcher with an unfounded omnipotence. One way to describe GPT-3 is &#8220;imagine a model trained by reading billions of words of internet text, and across books, and Wikipedia; it will have superhuman intuitions about all kinds of different topics&#8221;, though this is clearly incorrect. The GPT-3 learning algorithm clearly isn&#8217;t that sample efficient nor is the degree of generalisation as-good as this would imply. Superhuman AI researchers will certainly exist in the future, but to assume these early generations will have this kind of power seems to imply extremely radical improvements to the learning algorithm which do not strike us as easily gained.</p><h4>The productivity impact of the AI scientist could be capped, if the job of an AI researcher is to make &#8216;shot calls&#8217; about which experiments need more compute.</h4><p>One of the challenges of AI research is that experiments at different increments of scale can show very different performance. We&#8217;ve had RNNs, CNNs, and LSTMs (different neural network architectures for decades) but we&#8217;ve only have the computing power since the late 2000&#8217;s / early 2010s to be able to make use of them. In the episode of the Dwarkesh Podcast earlier cited, <a href="https://www.dwarkeshpatel.com/p/sholto-douglas-trenton-bricken">Sholto comments</a>:</p><blockquote><p>&#8220;[Y]ou never actually know if the trend will hold. For certain architectures the trend has held really well. And for certain changes, it's held really well. But that isn't always the case. And things which can help at smaller scales can actually hurt at larger scales. You have to make guesses based on what the trend lines look like and based on your intuitive feeling of what&#8217;s actually something that's going to matter, particularly for those which help with the small scale.&#8221;</p></blockquote><p>One way to think about what this does to your comment budget is that it divides it into a convergent series (or put another way, a series of Russian Dolls), up to the largest training run. We can do many small-scale experiments, increasing the scale until reaching the largest (and longest) training run. <a href="https://www.dwarkeshpatel.com/p/sholto-douglas-trenton-bricken">From the episode</a>;</p><blockquote><p>&#8220;Many people have a long list of ideas that they want to try, but paring that down and shot calling, under very imperfect information, what are the right ideas to explore further is really hard.&#8221;</p></blockquote><p>The biggest decisions the labs will make are &#8216;shot-calling&#8217; about what goes into the largest scales. I think there are two important questions here:</p><ol><li><p>To what extent are the potential gains to better decision making here capped?</p></li><li><p>To what extent are there marginal returns to intelligence for &#8216;shot calling&#8217;?</p></li></ol><p>It seems to us that some of the world&#8217;s smartest people, with decades of AI research experience making these calls; and it seems doubtful as to how much more <em>correctly </em>it is possible to make these decisions. We could be quite close to the limits of possible correctness in this task, for which marginal intelligence would not assist much.</p><h4>However, AI progress could go much more quickly than our picture suggests, if the digital AI scientists are much better at predicting the results of experiments than human researchers, or run at very low inference costs.</h4><p>Being able to predict the results of an experiment&#8212;what in humans we might summarise as &#8216;intuition&#8217;&#8212;is immune to compute bottlenecks. If the digital AI researchers are able to get superhuman intuition about which experiments to run, and which aren&#8217;t worth running, it could make the total research output rise dramatically, even whilst experimental throughput remains constant.</p><p>Furthermore, making a big inference efficiency improvement could dramatically improve the usefulness of AI researchers&#8212;a 50% gain would either mean the same &#8216;population&#8217; can run on half as much compute, with the other half going to experiments, or it is possible to run double the number of AI researchers. We&#8217;ve heard conflicting stories about whether this is possible. Some people have suggested that we&#8217;ve reached a &#8216;global minimum for inference costs&#8217;, as we will scale Chain of Thought reasoning faster than it is possible to make inference efficiency gains. On the contrary, others have been relatively optimistic about our capacity to make inference gains.</p><h3>How much can chip supply be scaled, if it needs to?</h3><p>This depends on the time scale.</p><p>The principal consideration here is frontier fab capacity&#8212;the H100 is fabricated using TSMC&#8217;s specialised process for AI chips called 4P (confusingly at <a href="https://www.techpowerup.com/gpu-specs/h100-sxm5-80-gb.c3900">5nm</a>), and the soon-coming Blackwell chips will use the same.</p><p>At the moment, AI accelerators make up a small fraction of 5nm capacity. For an estimate of how much they use, from <a href="https://www.ft.com/content/e85e43d1-5ce4-4531-94f1-9e9c1c5b4ff1">this FT article</a>, Microsoft bought 485,000 Hoppers &#8212; we will assume all H100s for simplicity &#8212; in 2024; Meta bought 224,000; Amazon, 196,000; Google, 169,000 (though of course, Google also has TPUs); and although it isn&#8217;t included, let&#8217;s assume that x.ai bought 125k GPUs. This is just short of 1.2 million H100 GPUs last year. Using the die size and some reasonable assumptions for yield, we can estimate this would have required 16,000 to 20,000 wafers. This is approximately a mere 0.3-0.4% of TSMC&#8217;s annual capacity in 5nm and below.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-22" href="#footnote-22" target="_self">22</a></p><p>Remember this is not the complete picture of demand&#8212;<a href="https://mohitdagarwal.substack.com/p/from-dominance-to-dilemma-nvidia">roughly half of NVIDIAs revenue comes from the hyperscalers</a>, the remaining half from neoclouds, startups, governments and so forth. Amazon, Microsoft, OpenAI are all developing their own custom silicon; and Google is on their 6th generation TPU. AMD is also making the MI325X to compete with NVIDIA&#8217;s next generation Blackwell. Furthermore, growth rates are high&#8212;NVIDIA&#8217;s most recent earnings reported revenue from their datacentre business (i.e. selling GPUs with all the extras to make them work) at <a href="https://nvidianews.nvidia.com/news/nvidia-announces-financial-results-for-third-quarter-fiscal-2025">$30.8 billion</a>. A year earlier, this was $14.51 billion; and a year before that it was $2.94 billion. In their <a href="https://nvidianews.nvidia.com/_gallery/download_pdf/664e53fe3d63322459a5eff6/#:~:text=NVIDIA%20(NASDAQ%3A%20NVDA)%20today,629%25%20from%20a%20year%20ago.">Q1 2024 earnings call</a>, TSMC gave guidance that they expect their AI processor business to grow at 50% per year, and become more than 20% of their revenue by 2028.</p><p>The production of AI accelerators could use capacity at 3nm and 4nm, but prices would need to rise, to outbid 3nm and 4nm demand, as there are higher costs for TSMC at the leading-edge.</p><p>There are potential bottlenecks in advanced packaging (CoWoS packaging capacity is <a href="https://www.trendforce.com/news/2024/09/05/news-tsmc-plans-rapid-cowos-expansion-through-2026-in-response-to-client-demand/#:~:text=The%20company%20expects%20CoWoS%20capacity,continuing%20at%20least%20until%202026.">lower than demand</a> until at least 2026) and in high-bandwidth memory, which limit scaling production in the next 1-2 years, though it is expected these bottlenecks will alleviate.</p><p>Building a leading-edge fab (chip factory) takes three to five years, though TSMC&#8217;s Arizona facility is delayed until 2027 or 2028 (construction began in December 2022) <a href="https://www.tomshardware.com/tech-industry/tsmc-delays-3nm-arizona-fab-by-a-year-cites-lack-of-us-subsidies-and-waning-demand#:~:text=Tech%20Industry-,TSMC%20delays%203nm%20Arizona%20fab%20by%20a%20year%2C%20cites%20lack,U.S.%20subsidies%20and%20waning%20demand&amp;text=TSMC%20faces%20another%20setback%20in%20the%20U.S.&amp;text=TSMC%2C%20the%20world's%20top%20foundry,the%20company%20said%20this%20week.">reportedly due</a> to a combination of low demand and uncertainty regarding US subsidies.</p><p>So, in short, there is a lot of room to grow within the existing fab capacity, though for the next couple of years this is limited by HBM and advanced packaging capacity. New fab construction requires significant capital expenditure, and therefore certainty of demand. New fabs could be commissioned to meet AI-related demand in future, but current projected demand in the 2020s remains far off this being the case.</p><p><em>We think the capacity to scale chip production is one of the most important inputs into the rate of progress, and so we will be dedicating a full piece to it, in the next edition of Inference.</em></p><h2>Expect big improvements in human welfare from AI automating science, but don&#8217;t expect that these gains will come quickly.</h2><p>Most of the improvements to human wellbeing in &#8216;frontier&#8217; economies comes from making scientific discoveries and turning these into new technologies. When AI lab leaders speak about the opportunity of AI, it is principally the scientific opportunity which they see as most exciting. Our scientific ambitions for AI should be enormous: ending disease, extending life, making abundant clean energy, more performant and green materials, ending unpleasant labour through robotic automation. We will also come to better understand human minds, wellbeing, and the most fundamental scientific questions. Like with AI research, automating R&amp;D depends on automating hypothesis generation, experimental design and implementation, and data analysis. Also like AI research, experimental throughput constraints scientific progress.</p><h3>Current AI systems can improve human researchers&#8217; hypothesis generation.</h3><p>The literature review process can be greatly enhanced with AI. <a href="https://www.futurehouse.org/">FutureHouse</a>, a non-profit research organisation, has built <a href="https://www.futurehouse.org/research-announcements/wikicrow">PaperQA2</a>, a literature review agent. Against a test of questions, where the answer was to be found only in the body of a single scientific paper, this agent was able to correctly answer 60.3% of questions. It was able to write cited, Wikipedia-style summaries which experts ranked as more accurate than the existing human-written Wikipedia articles in some scientific domains. It is notable that this agent was released in September 2024, before o1 or o3 were released, which will perform much better on GPQA (a benchmark of scientific expertise). It seems reasonable to imagine that by the middle of 2025, it will be possible for AI systems to write a literature review to postgrad-level.</p><p>Another example of the possibilities comes from Professor Derya Unutmaz, at the Jackson Laboratory, who studies cancer immunotherapy who prompted the model to support with experimental design, and <a href="https://x.com/DeryaTR_/status/1865111388374601806?lang=en-GB">subsequently wrote</a>:</p><blockquote><p>"While o1-Preview and GPT-4o were able to generate some interesting ideas based on this concept, but they were mostly what I could also conceive though better [than] most PhD students. In contrast, o1-Pro came up with far more creative and innovative solutions that left me in awe!"</p></blockquote><p>Here is the <a href="https://x.com/DeryaTR_/status/1865111388374601806">full output</a>.</p><p><strong>As an aside, language models can easily automate grant applications. </strong>This is quite a trivial use for AI systems now, clearly not using the frontier of their capabilities&#8212;but the <a href="https://thefdp.org/wp-content/uploads/FDP-FWS-2018-Primary-Report.pdf">Faculty Workload Survey of 2018</a> from the USFDP which received responses from 11,167 Principal Investigators said administrative requirements took 44.3% of their time. If we automated grant applications, would scientific output double as a result of the saved time? On the other hand, decreasing the cost of information processing might lead scientific funding institutions to impose greater reporting requirements, eating away any potential gains.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-23" href="#footnote-23" target="_self">23</a></p><p>Notably, all of these automations are pretty easy! They do not require large capital expenditures and scientists can choose to integrate these tools &#8216;bottom up&#8217;, <a href="https://www.research.pitt.edu/sites/default/files/how_chatgpt_is_transforming_the_postdoc_experience.pdf">as 1/3rd of postdocs already have</a>.</p><h3>Academia is poorly configured to adopt AI.</h3><ul><li><p>Because automation will happen lab-by-lab, there are low returns to scale<strong>.</strong></p></li><li><p>Because academic salaries are lower than other equivalent levels of labour skill, there are weaker incentives to automate<strong>.</strong></p></li><li><p>There are high capital costs to robotic lab automation, for which the returns are uncertain.</p></li><li><p>Incentive mechanisms in academia, for example, towards grant funding for longer periods, and positions being tied to projects, could make it harder to pivot research to focus on new capabilities.</p></li></ul><p>However, we can create new AI-first scientific institutions through structures like <a href="https://www.futurehouse.org/">FutureHouse</a>, <a href="https://www.convergentresearch.org/">Convergent Research&#8217;s FRO</a>, or <a href="https://www.aria.org.uk/">ARIA</a>. One parallel to think about this is how electricity improved productivity, not just in electrifying processes, but in making the assembly line possible&#8212;in what ways might we completely reorganize the institutions of science (perhaps around models, datasets, or experimental automation) that allows us to capture the full upside? Perhaps these things have to be reimagined from the ground up.</p><h3>Scientific automation faces many intrinsic headwinds.</h3><ul><li><p><strong>Most sciences require &#8216;real world&#8217; experiments </strong>whereas AI research only needs experiments within the computer. It is possible that progress in AI research will result in improvements to improve our simulations&#8212;of the cell, of particle physics, and so forth; that can partially substitute for experimentation, but there will be limits to this.</p></li><li><p><strong>There is no &#8216;programming language&#8217; for recording experimental design, </strong>in the same way that in AI research the experiments are recorded <em>exactly as they were implemented </em>in computer memory. This means that training agents to do AI research is much easier, because the AI labs will have a very rich corpus of data on which to train, whereas for agents to plan biomedical researcher experiments, we lack such a corpus and instead train on non-standardised descriptions of methods in academic papers.</p></li><li><p><strong>A very small fraction of the total data which research labs could collect, is collected. </strong>Organising data collection otherwise would impose big constraints on the scientists&#8217; productivity. On the contrary, computers capture AI research by design<br><br>This property of experimental research means that AI labs will have a corpus of negative results that would never be published, whereas in biomedical sciences, these are not published and are unlikely to be recorded in a structured format.</p></li></ul><h3>Chip R&amp;D is especially susceptible to these challenges described.</h3><ul><li><p>The processes for chip production, as well as for R&amp;D are highly secretive. Only TSMC will be able to automate what they do directly.</p></li><li><p>Knowledge about how to do chip research is often tacit and master-apprentice.</p></li><li><p>The complexity is so incredibly high&#8212;there are so many steps in the process, many of them are multivariate problems, and nobody has good visibility of the process.</p></li><li><p>As mentioned above, lots of chip research requires large amounts of real-world experimental throughput.</p></li></ul><p>However, this argument could be incorrect in a couple of important ways.</p><p>First, <strong>there are some steps in the chip R&amp;D process which have enormous leverage over other steps. </strong>Ultimately a large fraction of the performance improvement comes down to whether it is possible to shrink the node size&#8212;between Hopper and Blackwell it wasn&#8217;t, and quite a lot of the additional performance of Blackwell comes from how it was possible to make the chip <em>bigger</em>. (Granted this is still an important jump forward!)</p><p>DeepMind have developed AlphaChip, which has partially automated aspects of chip floorplanning (arranging where components go on the chip). This has reduced the wirelength, important for communication speed, on the TPUv6 by <a href="https://www.ctol.digital/news/ai-architect-google-alphachip-revolutionizes-chip-design/">6.2%</a>; but is somewhat bounded in its capacity to produce more computing power. The important question to ask about the chip process is: where are the marginal returns to intelligence very high?</p><p>Second, <strong>there can be unintuitive substitutes in R&amp;D. </strong>If you had asked us to explain why the discovery of protein structure does not get automated in 2014, we would plausibly explained the difficulty of automating the process of <a href="https://en.wikipedia.org/wiki/X-ray_crystallography">x-ray crystallography</a> &#8212; it&#8217;s hard to do the purification, to grow the crystals and so forth. Of course, we would have been wrong! Not in the direct sense that getting robots to x ray-crystallography would be easy, but that it turned out with a sufficiently large dataset of previous examples of x-ray crystallography and some hard-coded understanding of bond angles, we can create an AI system &#8212; AlphaFold &#8212; which is able to completely bypass this. Perhaps <em>extremely intelligent systems </em>will be able to spot much more difficult &#8216;bypasses&#8217; and take advantage of them.</p><h3>Biomedical advances will be bottlenecked by experimental throughput, and social welfare improvements will be bottlenecked by regulatory approval.</h3><p>Dario Amodei&#8217;s essay, <em><a href="https://darioamodei.com/machines-of-loving-grace">Machines of Loving Grace</a>,</em> details the changes which he thinks could arise from &#8216;powerful AI&#8217;, the model which results from some period (in his view, near), after the AI scientist where we have &#8216;a country of geniuses in a datacentre&#8217;. This model can control lab robots or tell humans which experiments to run. He thinks that we might be able to increase the speed of biomedical research progress by 10 times, which would mean in 5 to 10 years, we can make progress like:</p><ul><li><p>Reliable prevention and treatment of all infectious diseases</p></li><li><p>Elimination of most cancer</p></li><li><p>Very effective prevention and effective cures for genetic disease.</p></li><li><p>Prevention of Alzheimer&#8217;s</p></li><li><p>Improved treatments for most other ailments (diabetes, heart disease, autoimmune diseases and more.</p></li><li><p>Biological freedom (improvement to birth control, fertility, management of weight etc)</p></li><li><p>Doubling of the human life span through therapeutics.</p></li></ul><p>What would it take to increase experimental throughput by a factor of 10?</p><p>There are roughly 146,000 medical scientists employed in the United States. This would mean that 1.4 million people would be needed (or perhaps slightly fewer, as perhaps the job is more focused on experiments) to increase our throughput by a factor of 10, if the quality of experimental ideas is held constant. These people would need new buildings to work in, which typically take a year or two to build; and there would be many other things that have to happen like training the scientists at least in basic procedures, in order to get this scale-up. </p><p>On the other hand, we might not need a direct ten times increase in experimental throughout, perhaps if the models improve the quality of our ideas and the data we&#8217;re able to generate from each experiment, we need fewer experiments, so fewer buildings and so forth.</p><p>There is also the question of regulatory approval for new therapeutics. In this regard, Dario is also optimistic. He <a href="https://darioamodei.com/machines-of-loving-grace">writes</a>:</p><blockquote><p>Although there is a lot of bureaucracy and slowdown associated with them, the truth is that a lot (though by no means all!) of their slowness ultimately derives from the need to rigorously evaluate drugs that barely work or ambiguously work. This is sadly true of most therapies today: the average cancer drug increases survival by a few months while having significant side effects that need to be carefully measured (there&#8217;s a similar story for Alzheimer&#8217;s drugs). This leads to huge studies (in order to achieve statistical power) and difficult tradeoffs which regulatory agencies generally aren&#8217;t great at making, again because of bureaucracy and the complexity of competing interests.</p><p>When something works really well, it goes much faster: there&#8217;s an accelerated approval track and the ease of approval is much greater when effect sizes are larger. mRNA vaccines for COVID were approved in 9 months&#8212;much faster than the usual pace. That said, even under these conditions clinical trials are still too slow&#8212;mRNA vaccines arguably <em><a href="https://www.1daysooner.org/">should</a></em><a href="https://www.1daysooner.org/"> have been approved in ~2 months</a>. But these kinds of delays (~1 year end-to-end for a drug) combined with massive parallelization and the need for some but not too much iteration (&#8220;a few tries&#8221;) are very compatible with radical transformation in 5-10 years. Even more optimistically, it is possible that <a href="https://www.sciencedirect.com/science/article/pii/S135964462400134X">AI-enabled biological science</a> will reduce the need for iteration in clinical trials by developing better animal and cell experimental models (or even simulations) that are more accurate in predicting what will happen in humans. This will be particularly important in developing drugs against the aging process, which plays out over decades and where we need a faster iteration loop.</p></blockquote><p>We believe this understates the extent to which getting drugs through the regulatory approval process is bottlenecked by more capable systems. Semaglutide seems like the kind of magnitude of discovery we would hope that AI systems are able to produce for us, but the patents were first filed in 2008, and only within the last 5 years have we begun to see its scale up. Similarly, the Malaria vaccine <a href="https://worksinprogress.co/issue/why-we-didnt-get-a-malaria-vaccine-sooner/#the-emergence-of-the-rts-s-vaccine">needed 23 years in clinical trials</a>.</p><p>Getting regulatory approval bodies to use AI systems for therapeutic approval seems difficult, and unlikely. Their incentives are towards <a href="https://marginalrevolution.com/marginalrevolution/2015/08/is-the-fda-too-conservative-or-too-aggressive.html">risk-avoidance</a>, they would need to trust the evaluations from the simulations, and re-configure their processes in order to integrate these systems. (It seems more likely that a new, parallel regulatory agency would be set up to have AI analysis capabilities built natively into their approval process.) Additionally, this change would require the public to trust new methods of approval.</p><h3>Robotics progress will be accelerated by the automated AI researcher. However, we are sceptical of the most aggressive models of robotics deployment.</h3><p>Without making progress in robotics, the AI systems remain stuck on computers. There&#8217;s still a lot of leverage for systems which aren&#8217;t physically embodied, as we will discuss, there are a large number of tasks which can be done &#8216;remotely&#8217;. However, without making progress in robotics, further growth would become bottlenecked on manual labour from humans&#8212;it would be a <a href="https://en.wikipedia.org/wiki/Baumol_effect">Baumol effect</a>. The cost of information processing would fall, and so physical tasks would rise as a relative fraction of the economy.</p><p>Robotics progress is bottlenecked on data for them to learn how to act. OpenAI closed down their robotics work in 2021, in order to focus on language modelling, where there was a lot more data because of the Internet! (Note that OpenAI&#8217;s robotics team restarted a few days ago.) With the digital AI researcher, there will be two advantages to robotics progress: first, progress in video modelling&#8212;improving upon OpenAI&#8217;s work on SORA and DeepMind&#8217;s work on <a href="https://deepmind.google/research/publications/60474/">Genie</a> and <a href="https://deepmind.google/technologies/veo/veo-2/">veo</a>&#8212;could provide much better simulations of the real world, to train the model. The AI researcher could apply general AI algorithm improvements from training better computer-based systems, and also make progress on breakthroughs in computer vision algorithms, specifically for the robots.</p><p>There are a number of related questions for how fast we might see general robotics make a larger contribution to the economy:</p><ul><li><p>How readily can the digital AI researcher generalise to become a robotics researcher?</p></li><li><p>How quickly can algorithms for robots be improved, so as they can match human level performance?</p></li><li><p>How much of their training can happen in simulation instead of the real world?</p></li><li><p>How much prototyping and testing needs to happen on the hardware? How much of this testing can happen in simulation vs in the real world?</p></li><li><p>What are the holdout tasks &#8212; say, like getting a robotic hand to match human dexterity &#8212; which are going to prevent complete automation for some period?</p></li><li><p>How long do the holdout tasks <em>hold out?</em></p></li><li><p>Where does manufacturing happen, to what extent does this require skilled humans for assembly? Can the number of humans with these skills be scaled, until the robots can handle their own assembly?</p></li><li><p>How much training data can robots provide for models, and how useful?</p></li></ul><p><em>We will be dedicating a full piece to the robotics scale-up, as it seems to be one of the most important components of the growth story for the coming decades, meriting further investigation.</em></p><h2>Cognitive labour will be automated before physical labour, and could be automated much more quickly than previous technological revolutions.</h2><p>State-of-the-art AI systems have not made a very large GDP-level impact yet.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-24" href="#footnote-24" target="_self">24</a> This might seem bizarre&#8212;imagine going back to 2017 and showing someone the capabilities of OpenAI&#8217;s o3 model (very good software engineering, graduate-level scientist, and, from other models, near-perfect test scores in undergrad humanities). Surely one would have expected some noticeable increase to our collective output! It would be reasonable to have imagined big changes to &#8216;knowledge work&#8217; already.</p><h3>There are a number of factors that prevent AI systems performing tasks they are cognitively capable of doing.</h3><p><em>Throughout<strong> </strong>this section, it is worthwhile to reflect on whether these constraints <strong>continue to be true</strong> as the models become more capable. In some cases, the challenges to deployment fall away as models get more capable.</em></p><p>Right now, <strong>implementing current AI systems requires building &#8216;scaffolding&#8217; to improve reliability and performance. </strong>Most AI deployment in the real world has been workflows; where large language models and tools (like a code window, or internet access) are orchestrated through predefined code paths. These paths are known as scaffolding.</p><p>This happens because, until recently, language models have not been trained to perform tasks over time, just to predict the next token. To get them to do tasks, it would be necessary to stitch many prompts to the model together. Furthermore, businesses like deterministic processes. With scaffolding, the models are kept &#8216;on track&#8217; and their performance improves.</p><p>Knowing how to build scaffoldings and deploy AI workflows is currently a rare, and valuable skill. However, <strong>we expect this to get much easier with greater intelligence. </strong>One way to think about the current situation is that the &#8216;intelligence&#8217; is balanced between being inside the model&#8217;s parameters, and encoded in the &#8216;scaffolding&#8217; of business logic. Over time, as the models get smarter, a greater share of the processing will happen inside the models and less needs to be encoded in the scaffolding. This will happen as the models are trained for agency, and other algorithmic improvements advance their ability to act in unfamiliar contexts.</p><p>Right now, interacting with the models feels like talking to someone who is very smart, but has no context at all. Over time, as their context length, and management of their context window and memory improves, the models will be able to understand environments (like the internal documents of a company, or a large codebase) much more easily.</p><p>In the limit, <strong>we should expect that orchestrating agents will be done by other agents. </strong>The final step of OpenAI&#8217;s AGI research agenda is &#8216;coordination&#8217;&#8212;getting different AI agents to work together on problems. As part of this, one agent will be able to create a smaller &#8216;sub-agent&#8217; which could be specialised to a particular task.</p><p>Right now, <strong>implementing current AI systems requires businesses to reconfigure their processes. </strong>State-of-the-art models are capable in some domains, but they are &#8216;unbalanced&#8217; and struggle at some tasks which are very easy for humans. Model deployers need to account for this, and engineer the scaffolding for workflows, to allow humans to complete tasks which models struggle with. As it has been put previously, <a href="https://x.com/matthewclifford/status/1834271090295644477">&#8216;there are no AI-shaped holes lying around&#8217;</a>.</p><p>Over time, <strong>the models will need less structure to fit into organisations. </strong>The AI labs will improve the interactivity of the models&#8212;it will be possible to interact through many modalities in a much more natural way, and proactivity will be trained into the models during agency training. This is the opposite of what happens when we interact with Chatbots&#8212;they are trying to say less, to save on generation costs!</p><p>However, <strong>in many contexts, the deployment of AI systems is bottlenecked by organisational politics, and not the capabilities of the models. </strong>Nabeel Qureshi has an <a href="https://nabeelqu.co/reflections-on-palantir">excellent essay</a> reflecting on his experience working at Palantir, in which he writes:</p><blockquote><p>&#8220;[O]ften what really gets in the way is organizational politics: a team, or group, controls a key data source, the reason for their existence is that they are the gatekeepers to that data source, and they typically justify their existence in a corporation by being the gatekeepers of that data source (and, often, providing analyses of that data). This politics can be a formidable obstacle to overcome, and in some cases led to hilarious outcomes &#8211; you&#8217;d have a company buying an 8-12 week pilot, and we&#8217;d spend all 8-12 weeks just getting data access, and the final week scrambling to have something to demo.&#8221;</p></blockquote><p>Current systems might speed up the final week of production, but don&#8217;t affect the first 11 weeks of this project!</p><p>As in this case, in some domains, <strong>customers will have a preference for interacting with a human. </strong>There is some preliminary evidence that <a href="https://arxiv.org/pdf/2412.10849">o1-preview surpasses general practitioners on the reasoning aspects of diagnostics</a>. The paper gave o1-preview summarised descriptions of symptoms in text boxes, and asked it to provide a diagnosis. Of course, this only captures a very small fraction of <em>what it means to be a doctor</em>! Doctors are performing a blend of tasks&#8212;asking questions to establish the symptoms, expressing care and empathy, tailoring their explanations to the patient so they can understand, and making diagnoses. While it is highly likely that AI systems will surpass human capabilities to make diagnoses, most people will retain a preference for experiencing healthcare in person rather than through an online interface.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-25" href="#footnote-25" target="_self">25</a> The tasks which make up the job will be changed, and the interpersonal factors will increase in relative importance.</p><p>Furthermore, <strong>labour organisations may to resist automation. </strong><a href="https://www.michaelwebb.co/webb_ai.pdf">This paper</a> used semantic analysis of patent filings to predict where on the skill distribution we should expect AI systems to have the first impact. It predicts &#8216;upper-middle&#8217; jobs, like being a doctor, lawyer, or software engineer will be most exposed to automation from AI. As Webb notes <a href="https://80000hours.org/podcast/episodes/michael-webb-ai-jobs-labour-market/">in this podcast interview</a>, doctors and lawyers are likely to be able to lobby for regulations which require a human to remain &#8216;in the loop&#8217; in places they could otherwise be automated. On the other hand, intersectoral mobility for jobs which typically happen in cities and in &#8216;knowledge work&#8217; is typically easier than, say, adjustment for coal miners who lived in places where coal mining was the only industry. Furthermore, there are typically lower levels of labour organisation in &#8216;knowledge&#8217; work, and so we should perhaps expect less coordinated action.</p><p>Both factors, <strong>the preference for human-produced services and labour organisations, could be diminished by more capable systems</strong>. In some domains, by far-and-away superhuman AI systems would incentivise people to switch from human-produced services (imagine AI produced films vs human-produced films) and would make it ever-more difficult for labour organisations to ignore the differences in performance between the systems (imagine noticing rare conditions earlier and more accurately than human doctors could ever manage). However, there are some places where human services augmented by AI will always be superior to just AI; and there will be some responses to automation that are difficult to predict (e.g. on the grounds of safety; autonomous cars currently have 90% fewer crashes than humans over a large sample of miles driven, but have not been deployed more widely). There is also an income effect from higher automation - increasing the demand for human interaction.</p><p>At present, <strong>humans need to be liable for AI systems. </strong>There is no legal framework (as yet!) to determine liability between model developers, &#8216;product&#8217; companies (which fine tune models on proprietary data), and users of the models. But even if there was, models are not capable of acting unsupervised for long periods, they lack the error-correction, reasoning, and coherence to be reliable. Managing AI systems will become easier over time, as these capabilities are trained into the models, and so plausibly a single human can manage a larger team of agents as they need less frequent input from supervisors. We will also develop better infrastructure for escalating potential errors to AI and human oversight systems, to avoid mistakes. In the limit, AI agents will become capable of running entire firms in some industries; and humans will control a holding company for these firms, taking legal, but not managerial responsibility.</p><p><strong>Regulation could prevent AI use and bottleneck deployment and productivity gains. </strong>We turn down technology currently, which has detrimental effects both on economic growth and social welfare. For example, <a href="https://pubs.acs.org/doi/10.1021/es3051197">this paper</a> estimates that the slowdown in nuclear power construction since Chernobyl, in favour of fossil fuels, has cost 400,000 to 7 million lives globally. Nuclear could also provide electricity at prices competitive with solar, but due to a scientifically inaccurate approach to radiation exposure and poorly configured regulatory apparatus, it has been regulated to near-infeasibility.</p><p>However, AI might increase jurisdictional choice for businesses to operate, as they become less dependent on labour, and increasingly dependent on capital (datacentres, robots, robotic labs, factories etc). When this happens, it might cause regulatory approaches to soften, in order to capture some AI growth.</p><h3>AI systems will be leveraged by humans, mostly not AIs running firms.</h3><p>What should we expect AI automation to look like, concretely, in the near term?</p><p>The first AI agents to be released in the first few months of 2025 will be designed to complete simple tasks on your computer. Over time, these agents will become capable of performing a wider range of tasks on a computer, and acting for longer time-horizons. (For a more complete explanation, see &#8216;AGI is an engineering problem&#8217; in this edition.) Anything which a human does just on a computer, an AI agent will straightforwardly be able to do in the next few years.</p><p>Coding agents will be especially important to track here. Mark Zuckerberg has <a href="https://www.youtube.com/watch?si=PyFHjL2Y8PEyHFvZ&amp;t=7687&amp;v=7k1ehaE0bdU&amp;feature=youtu.be">said that</a>:</p><blockquote><p>&#8220;Probably in 2025, we at Meta, as well as the other companies that are basically working on this, are going to have an AI that can effectively be a sort of midlevel engineer that you have at your company that can write code.&#8221;</p></blockquote><p>It is quite difficult to measure to what extent the actual productivity of software engineers is improved by this&#8212;lines of code does not mean more real-world output, as perhaps we bury ourselves in poorly designed software! However, it seems reasonable to imagine that software productivity could rise quite steeply.</p><p>We expect that knowledge workers will be managing groups of agents for computer-based tasks, using tools like <a href="https://www.microsoft.com/en-gb/microsoft-copilot/microsoft-copilot-studio">Microsoft Copilot Studio</a>, and over time, as the models&#8217; capacity for long-horizon tasks improves, and their reasoning and planning capabilities advance further, the number of agents each human can manage will expand. We expect this to be like: &#8220;all humans are getting promoted&#8221;, at least in the near term.</p><p>Lots of knowledge work bundles cognitive capabilities with &#8216;embodied&#8217; human skills, for example, product management, consulting, medical care, and law. We think humans will continue to do these jobs, because whilst the cognitive capabilities could be in-principle automated by AI; they are complemented by things which AIs will struggle to do, at least in the near term. Be charismatic, be empathetic, be &#8216;present&#8217;, be accountable. (Perhaps there is a vision for robotics whereby close mimicry of human emotion and embodiment is possible, though we exempt this from our analysis.)</p><p>There are a set of knowledge work professions which are not complemented by this kind of embodiment. For example, making good films, making software, and running a hedge fund. It is plausible that humans continue to do some aspect of this (telling the AI system what film to make, what program to write, or raising LP money for the AI running the hedge fund) though we think it is possible to automate these sectors entirely. If an AI hedge fund was likely to make better returns, you would invest with them, irrespective of the hedge fund sales.</p><p>We do not expect &#8216;AI composed firms&#8217; to make up a majority of sectors.</p><h2>Automating remote tasks could lead to rapid growth, but is very unlikely to lead to explosive growth.</h2><p>We can build an economic model of tasks in the economy, and estimate which fraction of these can be automated, to predict how much AI causes growth to accelerate, under our assumptions mentioned above. There are three ingredients to any model of this:</p><ol><li><p>What fraction of tasks are automated?</p></li><li><p>How well can these substitute for non-automated tasks?</p></li><li><p>How cheap are the AI systems that can substitute for human cognitive labour?</p></li><li><p>How much do Baumol and Engels effects make output growth concentrated in a small number of sectors less valuable?</p></li></ol><p>Below we list our parameter estimates, if you would like to skip to our conclusions you can do so here.</p><h3>What fraction of tasks are automated?</h3><p>For the purposes of this, we are going to assume that all tasks which can be done on a laptop are automatable, and all tasks which require physical presence are not automatable in this model.</p><p>In the ordinary economy, there are a number of tasks which are not done &#8216;remotely&#8217;, but could be. Whilst most firms prefer their software engineers to be in the office, it is possible to do software engineering remotely.</p><p>The pandemic is a useful case study to understand, in the limit, how many jobs could be done remotely if they needed to. This survey suggests that 37% of jobs were done remotely during the pandemic. This is likely to be an overestimate for the number of jobs which are <em>actually </em>automatable, because when people were sent home for lockdown the question was not <em>which jobs are suitably completed remotely</em>, but rather, <em>whether it is possible to get any output from workers who are stuck at home.</em></p><p>To adjust for this, we can break down <em>by task </em>instead of <em>by job</em>. <a href="https://epoch.ai/gradient-updates/consequences-of-automating-remote-work">Barnett 2025</a> uses GPT-4o to analyse O<em>NET</em>, the database of what workers in 1000 professions spend their time doing in the US economy; and finds that 34% of the tasks were automatable. This is quite different from 37% of jobs, because 34% of tasks can be spread across a wide range of workers in a way that could be bundled to affect production. For example, the scientist can write their grant proposal remotely, but if they cannot get access to the lab, their output will still be 0.</p><p>Estimating what fraction of tasks have an automatable (i.e. remote) component, bundled with an non-automatable (i.e. in person) component; is very difficult. To return to the earlier example of a Palantir engineer implementing a data platform, the cognitive task to figure out what product a company needs, is possible to do remotely, but it relies upon asking them questions to establish their needs, and building rapport&#8212;necessarily not automatable. To deal with this, and set a lower bound on the number of automatable tasks in the economy; we look at what fraction all of their top 5 tasks are remote. <a href="https://epoch.ai/gradient-updates/consequences-of-automating-remote-work">Barnett 2025</a> provides that this is 13%.</p><h3>How well can automated tasks substitute for non-automated tasks?</h3><p>Two reasons why task automation might be good</p><ol><li><p>Spend less money on it.</p></li><li><p>Because it is possible to do that task more relative to another task, if it can substitute for it. (if you have better software, you can manage your inventory better, which means lower inventory, so lower trolleys for moving things around.)</p></li></ol><p>We want to set a parameter (number between zero and infinity) which captures to what extent an automated task can do this substitution. How can we estimate this?</p><p>One of the ways <a href="https://epoch.ai/gradient-updates/consequences-of-automating-remote-work">Barnett 2025</a> estimates this is, again, by looking at the pandemic. At this point, there was a very large increase in the number of remote workers relative to in-person workers. This is an imperfect proxy for what happens when AI systems can be added to the economy quite easily (more remote workers). Total economic output didn&#8217;t decline very much&#8212;roughly 8%&#8212;which is surprisingly small, given that everyone was made to stay at home. This scenario produces an &#8216;elasticity of substitution&#8217; (the parameter for how much an automated task can substitute for a non-automatable task) of 13.</p><p>This is extremely large! However, we don&#8217;t place much weight on this for a couple of reasons:</p><ul><li><p>During the pandemic, roughly 1/3rd of US workers went remote. But, as we&#8217;ve already noted, 34% of tasks can be done from home. So all that has happened, roughly, is that these tasks have begun to happen from home. It does not, necessarily, tell us <em>how well </em>this 34% of tasks can substitute for the other 66% of tasks.</p></li><li><p>Remote work requires infrastructure. The pandemic prevents investment and maintenance from occurring. In the short run, this is not a big problem (it might even increase output); but in the long-run this problem becomes large. This can cause one to overstate the elasticity of substitution in the pandemic example.</p></li></ul><p>What is another approach to estimating this? Very roughly, we can say that higher skilled workers tend to do cognitive tasks, and lower-skilled workers tend to do physical tasks. (We appreciate this loses a lot of individual difference, but we make a simplification for modelling purposes.) So it is instructive to see to what extent higher skilled workers can substitute for lower skilled workers. This provides an <a href="https://direct.mit.edu/rest/article/106/5/1187/112420/Publication-and-Attenuation-Biases-in-Measuring">elasticity of 4</a>.</p><p>What are some reasons to be suspicious of this? In order for things to have an elasticity greater than one, they need to be substitutes. Simply, what substitutes are, is when you can still produce something, if you only have one of them. We claim that high and low skilled workers are an example of this. High skilled workers can plausibly perform the tasks which lower skilled workers can perform. (We appreciate there will be exceptions, and the idea of &#8216;lower skill&#8217; can be refuted, but again, this is a simplification for modelling purposes.) We would expect remote and in-person work to not be like this: if everyone was remote, then who would repair the factories, run the power stations, and care for the patients?<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-26" href="#footnote-26" target="_self">26</a> This means high and low skill labour have to give an elasticity of substitution above 1, but this does not mean that remote and in-person work will have the same relationship.</p><p>Following <a href="https://epoch.ai/gradient-updates/consequences-of-automating-remote-work">Barnett 2025</a>, we&#8217;ll use an elasticity of 0.5 as a lower bound.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-27" href="#footnote-27" target="_self">27</a></p><h3>How cheap are the AI systems that can substitute for human cognitive labour?</h3><p>We separately estimate the model at $4, $10, and $40/hour.</p><p>$10/hour is a very rough estimate of the cost for o1 tokens, generating for an hour; and $4/hour and $40/hour are &#8216;a bit above&#8217; and &#8216;a bit below&#8217;. We recognise more precision could be useful here, and also note that this assumes inference costs remain roughly aligned with o1. There are a wide range of views about what will happen to inference costs in the future. As we argued earlier in the section on <a href="https://docs.google.com/document/d/1mRzI1gYf_0Kmag23JMDrOlTCAl6r7TK4JynZZDzztXw/edit?tab=t.0#heading=h.idrzg3s4fl0y">the slow rate of hardware commoditisation</a>, Chain of Thought inference costs could well rise nonlinearly, although inference efficiency improvements will push in the opposite direction.</p><h3>How much do more workers help growth?</h3><p><a href="https://epoch.ai/gradient-updates/consequences-of-automating-remote-work">Barnett 2025</a> assumes constant returns to scale in adding more workers&#8212;simply, twice as many workers leads to twice as much output, forever. The problem is that, in fact, workers need capital to produce things (e.g. machines, computer, tools etc to work with). If there are a lot more workers, without investing to raise the capital stock, the increases in output are determined by the elasticity of substitution between capital and labour. (Stepping back from the economics jargon,) we can either spend resources on adding a marginal remote worker&#8212;in this case an AI agent&#8212;or we can invest in more machines or tools&#8212;to make our existing AI agents (and humans) more productive&#8212;or we can spend those resources on consumption&#8212;the whole reason we&#8217;re doing this anyway!<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-28" href="#footnote-28" target="_self">28</a></p><p>At this point, a sensible rebuttal could be, &#8220;Don&#8217;t labour and capital in this case both mean datacentre compute?&#8221; This would be correct in sectors which are completely automated, but we&#8217;re interested in what happens after this (these few sectors which are completely automated, by themselves, do not set off explosive growth). For example, doctors can benefit from more AI assistance interpreting scans, but if they want more scans, they&#8217;ll also need more MRI machines. Likewise, physicists can do many more simulations of experiments in a datacentre, but ultimately if they want to test them they&#8217;ll still need bigger particle accelerators.</p><p><a href="https://arxiv.org/abs/2309.11690">Erdil &amp; Besiroglu 2023</a> uses a literature review of estimates of the capital-labour elasticity of substitution, and this gives <a href="https://onlinelibrary.wiley.com/doi/full/10.1111/obes.12312?casa_token=f_y-uouRE-wAAAAA%3A5US4G_NKbsbrKZM3JyezuJRBhkYBaoWqQcLR8qS4ePTbvedGlQl0w2j0hCPRBVDx6LR1_gc33O83Xw">0.45-0.87</a>. However, in the literature review they cite, there is a distortion of the distribution of papers investigating capital-labour elasticity, whereby if people get a negative value, they do not publish their results. A negative value would mean that production would have been higher if there was less labour and less capital, and so people choose not to publish. In the graphs below, which show the distribution of capital-labour elasticities in the studies, there is a clear truncation of the left tail.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!h-9-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6125cc5-2c0d-4c3e-8465-fc2e94c6b2e3_729x288.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!h-9-!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6125cc5-2c0d-4c3e-8465-fc2e94c6b2e3_729x288.png 424w, https://substackcdn.com/image/fetch/$s_!h-9-!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6125cc5-2c0d-4c3e-8465-fc2e94c6b2e3_729x288.png 848w, https://substackcdn.com/image/fetch/$s_!h-9-!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6125cc5-2c0d-4c3e-8465-fc2e94c6b2e3_729x288.png 1272w, https://substackcdn.com/image/fetch/$s_!h-9-!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6125cc5-2c0d-4c3e-8465-fc2e94c6b2e3_729x288.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!h-9-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6125cc5-2c0d-4c3e-8465-fc2e94c6b2e3_729x288.png" width="729" height="288" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a6125cc5-2c0d-4c3e-8465-fc2e94c6b2e3_729x288.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:288,&quot;width&quot;:729,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!h-9-!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6125cc5-2c0d-4c3e-8465-fc2e94c6b2e3_729x288.png 424w, https://substackcdn.com/image/fetch/$s_!h-9-!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6125cc5-2c0d-4c3e-8465-fc2e94c6b2e3_729x288.png 848w, https://substackcdn.com/image/fetch/$s_!h-9-!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6125cc5-2c0d-4c3e-8465-fc2e94c6b2e3_729x288.png 1272w, https://substackcdn.com/image/fetch/$s_!h-9-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa6125cc5-2c0d-4c3e-8465-fc2e94c6b2e3_729x288.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-29" href="#footnote-29" target="_self">29</a></p><p>Therefore, the elasticity of labour-capital substitution is overestimated in Erdil &amp; Besiroglu 2023. <a href="https://www.sciencedirect.com/science/article/pii/S1094202521000387?via%3Dihub">Gechert et al 2022</a> corrects this, and provides a capital-labour elasticity of 0.3.</p><p>This implies it is quite difficult to substitute labour for capital, and as such, even with a very large increase in workers, growth is bottlenecked by capital. This capital can be accumulated&#8212;although, due to imperfect insurance markets and finitely lived agents, <a href="https://doi.org/10.3982/ECTA19417">potentially not fully</a>. Getting through this will be slow and expensive.</p><h4>What does this understanding imply for economic growth</h4><p>We will now set out three cases for growth. Note that all figures are additional growth over whatever exogenous technological progress occurs.</p><p>The maximally optimistic case.</p><ul><li><p>Elasticity of substitution = 4</p></li><li><p>% of tasks automatable = 35</p></li><li><p>Agents cost $4/hour</p></li></ul><p>Realistic (upper bound)</p><ul><li><p>Elasticity of substitution = 2</p></li><li><p>% of tasks automatable = 30</p></li><li><p>Agents cost $10/hour</p></li></ul><p>Lower bound</p><ul><li><p>Elasticity of substitution = 0.5</p></li><li><p>% of tasks automatable = 20</p></li><li><p>Agents cost $40/hour</p></li></ul><p>In the maximally optimistic case, output grows at 19.3% per year on average over the next 20 years, and 16.3% per year over the next 100 years, although we think this is unlikely under purely cognitive automation for the reasons outlined above.</p><p>In our upper bound, output grows by 12.2% per year on average over the next 20 years, and 6.8% in the next 100 years.</p><p>In our lower bound, output grows by 3.5% per year on average over the next 20 years, and 0.7% per year over the next 100 years.</p><p>But, now we need to introduce <a href="https://en.wikipedia.org/wiki/Baumol_effect">Baumol effects</a>. If automation is better in some industries than others, say films, finance, and programming; there are diminishing returns to having ever-greater amounts of B2B SaaS in the world, and so as a result, prices will fall in the industries where we are producing lots of goods, and so those industries share of GDP will rise by less than expected. Because of how GDP is constructed, at least half of the growth that would have occurred, less Baumol effects, will occur but this still dampens output by a lot.</p><p>But now we need to introduce <a href="https://doi.org/10.3982/ECTA15202">Engel effects</a>. As automation increases productivity and income, people tend to shift their spending toward sectors like healthcare, education, and housing - areas that historically show slower productivity growth. This phenomenon, known as Engel effects, means that as societies become wealthier, an increasing share of spending goes to these slower-growing sectors. In the US over the past few decades, Engel and Baumol effects have reduced output growth by approximately <a href="https://doi.org/10.3982/ECTA15202">25%</a> (not 25pp) compared to what it would be otherwise.</p><p>Taking these Engel effects into account, we estimate that economic growth will be 3% to 9% higher per year for the 20 years following significant AI automation. This range reflects both the potential for productivity gains in highly automatable sectors and the dampening effect of spending shifts toward sectors where automation gains may be more limited.</p><h1>Conclusion</h1><p>This quote from <em>The Optimistic Thought Experiment </em>on the nature of the dotcom bubble, provides a useful lens for thinking about the economic impact of AI (albeit if the 'doom&#8217; perspective is a little strong).</p><blockquote><p>It is often claimed that the mass delusion reached its peak in March 2000; but what if the opposite also were true, and this was in certain respects a peak of clarity? Perhaps with unprecedented clarity, at the market &#8217;s peak investors and employees could see the farthest: They perceived that in the long run the Old Economy was surely doomed and believed that the New Economy, no matter what the risks, represented the only chance.</p></blockquote><p>The fundamental thesis&#8212;that AI research output will be automated; that humanity will create &#8216;superintelligent&#8217; systems; and that AI systems will do science that create greater and faster technological progress than humans could ever have done&#8212;will be borne out in the fullness of time. But this vision has to make contact with reality, and reality can act as a weird breaking mechanism: Meta wants to build AGI, but <a href="https://www.ft.com/content/ed602e09-6c40-4979-aff9-7453ee28406a">they couldn&#8217;t use a nuclear power plant for their datacentre, because of some rare bees</a>.</p><p>These bits never make it in the sci-fi novels, and so it&#8217;s easy to see far into the future, but miss the (frankly bizarre) hurdles along the way.</p><p>AI will still be constrained by physical and institutional bottlenecks&#8212;drug development still requires clinical trials, chip fabs take years to build, lab experiments need physical space and technicians, and negotiations between people need to take place. Lots of cognitive tasks will be automated, in our view, quite quickly, but because of the requirements for human liability and often necessary complements between cognitive tasks and &#8216;embodied&#8217; tasks, we anticipate knowledge workers will be augmented, not replaced. In some sense, it could feel like &#8216;a promotion for everyone&#8217;.</p><p>Over time, the growth effects will rise beyond our current view, but accounting for present bottlenecks we expect an annual growth boost from AI to be 3-9%&#8212;transformative, but not explosive.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Many thanks to Byrne Hobart, Mike Webb, Dan Carey, Jason Hausenloy, Phil Trammel, Oliver Jaffe, Olivia Benoit and Eduard Baryon for invaluable feedback and comments in the process of writing this piece.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Jovanovic &amp; Rousseau (2005), General-Purpose Technologies</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>Crafts (2004), Steam as a General Purpose Technology</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>David (1991), The Dynamo and the Computer</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>Davidson (2023), <a href="https://www.openphilanthropy.org/research/what-a-compute-centric-framework-says-about-takeoff-speeds/">What a Compute-Centric Framework Says About Takeoff Speeds</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>Erdil and Besiroglu (2024), Explosive growth from AI automation: A review of the arguments, Barnett (2025), The economic consequences of automating remote work, Hanson (1997), Economic Growth Given Machine Intelligence</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>Trammell and Korinek (2023), Economic growth under transformative AI</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>We don&#8217;t intend to suggest there is a single, monolithic view at AI labs, nor that everyone there will subscribe to the strongest version of this view, but we mean to sketch a perspective on the dominant intellectual paradigm which a large proportion of actors are operating in.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p>&#8220;For the rationalists of the eighteenth and nineteenth centuries, as well as for all those who consider themselves cosmopolitan today, this sort of hysterical talk about the end of the world was deemed to be the exclusive province of people who were either stupid or wicked or insane (although mostly just stupid). Scientific inculcation would replace religious indoctrination. Today, we no longer believe that Zeus will strike down errant humans with thunderbolts, and so we also can rest peacefully in the certain knowledge that there exists no god who will destroy the whole world.<br><br>And yet, if the truth were to be told, our slumber is not as peaceful as it once was. Beginning with the Great War in 1914, and accelerating after 1945, there has re-emerged an apocalyptic dimension to the modern world. In a strange way, however, this apocalyptic dimension has arisen from the very place that was meant to liberate us from antediluvian fears. This time around, in the year 2008, the end of the world is predicted by scientists and technologists.&#8221; &#8212; <a href="https://www.hoover.org/research/optimistic-thought-experiment">The Optimistic Thought Experiment</a>, January 2008</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-10" href="#footnote-anchor-10" class="footnote-number" contenteditable="false" target="_self">10</a><div class="footnote-content"><p>However, bank tellers <a href="https://www.bitsaboutmoney.com/archive/why-is-that-bank-branch-there/">did decrease</a> both as a fraction of bank employment and per branch.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-11" href="#footnote-anchor-11" class="footnote-number" contenteditable="false" target="_self">11</a><div class="footnote-content"><p>These are known as the <a href="https://en.wikipedia.org/wiki/Consumer_choice#Price_effect_as_the_sum_of_substitution_and_income_effects">income effect and substitution effect</a>, respectively.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-12" href="#footnote-anchor-12" class="footnote-number" contenteditable="false" target="_self">12</a><div class="footnote-content"><p><a href="https://epoch.ai/files/Interviewing_AI_researchers_on_automation_of_AI_R_D.pdf">Automation of AI R&amp;D: Researcher Perspectives</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-13" href="#footnote-anchor-13" class="footnote-number" contenteditable="false" target="_self">13</a><div class="footnote-content"><p>Again, there could be many more here! These benchmarks, in the words of their creators, could be greatly improved upon! Please consider working on this.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-14" href="#footnote-anchor-14" class="footnote-number" contenteditable="false" target="_self">14</a><div class="footnote-content"><p>While these tests are at a smaller sample, the solutions are not going to be found online (as may be the case with Kaggle) so the models have not seen these problems before, these problems are likely to be more reflective of what ML engineering looks like for researchers in AI labs, and there is more control over the environment which humans perform the tests in. There are benefits to both approaches.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-15" href="#footnote-anchor-15" class="footnote-number" contenteditable="false" target="_self">15</a><div class="footnote-content"><p><a href="https://metr.org/blog/2024-11-22-evaluating-r-d-capabilities-of-llms/">Evaluating frontier AI R&amp;D capabilities of language model agents against human experts</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-16" href="#footnote-anchor-16" class="footnote-number" contenteditable="false" target="_self">16</a><div class="footnote-content"><p>Optimising a kernel for latency refers to modifying the core program code that interfaces between computer hardware and software to minimize the time delay between when an instruction is initiated and when it is completed, often involving careful tuning of memory access patterns, thread scheduling, and hardware-specific optimizations.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-17" href="#footnote-anchor-17" class="footnote-number" contenteditable="false" target="_self">17</a><div class="footnote-content"><p>AI labs should be very sensitive to reputation, e.g. nobody wants to be known as the place where people work after Anthropic or Google or OAI rejected them. Whereas if your lab has a GPU shortage, it's fine&#8212;Nvidia doesn't make as many as the market will bear, and the academics you're hiring are used to an even more acute shortage, so your pitch to them is something like "I'm sorry to say you may be so unlucky as to only have 100x the resources you have right now. Wish we could do better!"</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-18" href="#footnote-anchor-18" class="footnote-number" contenteditable="false" target="_self">18</a><div class="footnote-content"><p><a href="https://ifp.org/future-of-ai-compute/">How to Build the Future of AI in the United States</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-19" href="#footnote-anchor-19" class="footnote-number" contenteditable="false" target="_self">19</a><div class="footnote-content"><p><a href="https://semianalysis.com/2024/04/10/nvidia-blackwell-perf-tco-analysis/">Nvidia Blackwell Perf TCO Analysis &#8211; B100 vs B200 vs GB200 NVL72</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-20" href="#footnote-anchor-20" class="footnote-number" contenteditable="false" target="_self">20</a><div class="footnote-content"><p>https://x.com/eladgil/status/1827521805755806107</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-21" href="#footnote-anchor-21" class="footnote-number" contenteditable="false" target="_self">21</a><div class="footnote-content"><p>Tokens have awful margins, and are impossible to price discriminate on. However, moving to other industries may be hard.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-22" href="#footnote-anchor-22" class="footnote-number" contenteditable="false" target="_self">22</a><div class="footnote-content"><p>Our estimates are in line with Epoch AI&#8217;s estimate that &#8220;2 million H100 GPUs [their projected total demand for 2024] would consume only 5% of the 5nm node capacity&#8221;. The differences arise from our inclusion of 4nm and 3nm capacity, and our focus on just hyperscale customers.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-23" href="#footnote-anchor-23" class="footnote-number" contenteditable="false" target="_self">23</a><div class="footnote-content"><p>Funding institutions could also respond in other ways - such as a greater reliance on personal networks, automated application assessment, etc in ways which change the funding landscape in other ways with ambiguous effects on output.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-24" href="#footnote-anchor-24" class="footnote-number" contenteditable="false" target="_self">24</a><div class="footnote-content"><p>US GDP has been growing relatively well over the past 2 years, with 5 consecutive quarters of 0.5% productivity growth in the US vs only 2 quarters over all of 2015-2019 with this, a development which may be related to AI but is hard to disentangle from other factors at present.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-25" href="#footnote-anchor-25" class="footnote-number" contenteditable="false" target="_self">25</a><div class="footnote-content"><p>This should be celebrated as an unequivocal good for the world&#8212;people will be able to consume much more healthcare, unlimited by the scarcity of appointments and long wait times if they are willing to consume online; and access to the world&#8217;s best medical care is democratised to everyone in the world.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-26" href="#footnote-anchor-26" class="footnote-number" contenteditable="false" target="_self">26</a><div class="footnote-content"><p>However, it has also been argued by some that advanced AI might succeed in effectively massively increasing worker skill levels by giving everyone VR access to inform the person what to do with their hands. This wouldn&#8217;t work with areas requiring trained dexterity (e.g. being a surgeon, violinist or top chef) but it could result in a reasonably high elasticity of substitution, at least for a while.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-27" href="#footnote-anchor-27" class="footnote-number" contenteditable="false" target="_self">27</a><div class="footnote-content"><p>The paper he cites for the elasticity of 0.5 seems to be estimating an income elasticity rather than an elasticity of substitution, but we will follow.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-28" href="#footnote-anchor-28" class="footnote-number" contenteditable="false" target="_self">28</a><div class="footnote-content"><p>There is one other potential use of these agents - intensifying research output. However, research plausibly faces the same problems regarding diminishing marginal returns to labour with the same capital stock. Additionally, <a href="https://web.stanford.edu/~chadj/IdeaPF.pdf">ideas seem to be getting much harder to find</a> over time, dampening the increase in research output from investing more resources in the sector further. In our calibration, a substantial majority of the gains from AGI went to lowering research intensity rather than raising output, hence we omitted it from the model as the effect on the results was small. </p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-29" href="#footnote-anchor-29" class="footnote-number" contenteditable="false" target="_self">29</a><div class="footnote-content"><p><a href="https://wiley.scienceconnect.io/api/oauth/authorize?ui_locales=en&amp;scope=affiliations+alm_identity_ids+login_method+merged_users+openid+settings&amp;response_type=code&amp;redirect_uri=https%3A%2F%2Fonlinelibrary.wiley.com%2Faction%2FoidcCallback%3FidpCode%3Dconnect&amp;state=MD7fwcD8HXzc3nKJa2pRerTAMbDaB9ZGZdqCdr0pZaTrxD5x19vlmLTAMbDaB9ZGZdqCdr0pZaR6V%2BommeWPxrTAMbDaB9ZGZdqCdr0pZaR3qyEpGB0h2mnYKk2egCfX&amp;prompt=none&amp;nonce=QYj7V%2BMbAK%2B229CW%2BKAlUl7QEu1FqQmL6qwQPkntwws%3D&amp;client_id=wiley">The Elasticity of Substitution Between Capital and Labour in the US Economy: A Meta-Regression Analysis</a></p></div></div>]]></content:encoded></item><item><title><![CDATA[AGI is an Engineering Problem]]></title><description><![CDATA[Until this decade, artificial general intelligence was a scientific problem.]]></description><link>https://inferencemagazine.substack.com/p/agi-is-an-engineering-problem</link><guid isPermaLink="false">https://inferencemagazine.substack.com/p/agi-is-an-engineering-problem</guid><dc:creator><![CDATA[Jason Hausenloy]]></dc:creator><pubDate>Fri, 17 Jan 2025 17:22:39 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!6CrP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e81eff1-a062-40b8-a05c-e9f476171c99_1284x534.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Until this decade, artificial general intelligence was a scientific problem. <strong>The main ideas to build it were </strong><em><strong>missing.</strong> </em>In 1999, Shane Legg (cofounder of Google DeepMind), <a href="http://www.vetta.org/2009/12/tick-tock-tick-tock-bing/">predicted</a> we&#8217;d build AGI in 2028 based on extrapolations of compute power trends. His prescience on reinforcement learning is remarkable, but the vision was necessarily fuzzy. <strong>This is no longer the case. </strong>Sam Altman <a href="https://blog.samaltman.com/reflections">announced</a> recently:</p><blockquote><p>We are now confident we know how to build AGI as we have traditionally understood it...[w]e are beginning to turn our aim beyond that, to superintelligence in the true sense of the word.</p></blockquote><p>Building AGI has become an engineering problem.</p><div><hr></div><h3>The &#8216;Eye of Sauron&#8217;-Theory-of-Research-Progress</h3><p>When thinking about future AI progress, follow the research priorities of leading AI labs. I often imagine their research focus as &#8220;The Eye of Sauron&#8221;, the great flaming eye from Lord of the Rings. What The Eye gazes upon becomes the industry's all-consuming focus; what it ignores remains unsolved - not because it's impossible, but because it's not yet time.</p><p>Take emotional intelligence. Perhaps the model needs to sound more natural, needs lower latency, needs to reason about the information you give it, textually, from your voice, and from body language. Or perhaps, it just needs to text you first. Increased latency, personality post-training and better UIs gets most of the way there. But The Eye isn&#8217;t looking here yet.</p><p>For the past few years, The Eye has settled its gaze squarely at scaling training. And instead of trying to fully elicit the capabilities of the model, or fix particular quirks in its behaviour, the attitude has been &#8220;just scale&#8221;, and it&#8217;ll be fixed.</p><h3>Listen to the labs</h3><p>To see where The Eye points next, we don't have to guess - we can listen to what the labs have told us. OpenAI's &#8220;five levels of AGI&#8221;, first shown in an employee all-hands and later <a href="https://www.bloomberg.com/news/articles/2024-07-11/openai-sets-levels-to-track-progress-toward-superintelligent-ai">published</a> by Bloomberg, maps out their critical path: first chatbots, then reasoners, next agents; before 'innovators', and finally 'organisations'.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!6CrP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e81eff1-a062-40b8-a05c-e9f476171c99_1284x534.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6CrP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e81eff1-a062-40b8-a05c-e9f476171c99_1284x534.png 424w, https://substackcdn.com/image/fetch/$s_!6CrP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e81eff1-a062-40b8-a05c-e9f476171c99_1284x534.png 848w, https://substackcdn.com/image/fetch/$s_!6CrP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e81eff1-a062-40b8-a05c-e9f476171c99_1284x534.png 1272w, https://substackcdn.com/image/fetch/$s_!6CrP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e81eff1-a062-40b8-a05c-e9f476171c99_1284x534.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6CrP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e81eff1-a062-40b8-a05c-e9f476171c99_1284x534.png" width="1284" height="534" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8e81eff1-a062-40b8-a05c-e9f476171c99_1284x534.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:534,&quot;width&quot;:1284,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!6CrP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e81eff1-a062-40b8-a05c-e9f476171c99_1284x534.png 424w, https://substackcdn.com/image/fetch/$s_!6CrP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e81eff1-a062-40b8-a05c-e9f476171c99_1284x534.png 848w, https://substackcdn.com/image/fetch/$s_!6CrP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e81eff1-a062-40b8-a05c-e9f476171c99_1284x534.png 1272w, https://substackcdn.com/image/fetch/$s_!6CrP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e81eff1-a062-40b8-a05c-e9f476171c99_1284x534.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Why these levels? The goal of the AGI companies is to have organisations of automated AI researchers, so that increasing the scale of training runs can be improved at an ever improving rate. The AI researchers would be able to work on any capability that is required to meet any of many definitions of AGI today, and beyond.</p><p>Understanding this critical path is useful in two ways:</p><p>First, it <strong>helps one track the most important events in AI and understand their true significance in solving &#8220;what&#8217;s left&#8221; to create powerful systems</strong>. For example, most of Google DeepMind's early work on beating video games like Atari wasn't because they cared about gaming per se, but about solving key problems in planning and goal-directedness (and their <a href="https://deepmind.google/discover/blog/genie-2-a-large-scale-foundation-world-model/">recent</a> foundation world model, which can generate interactive environments, are explicitly for training agents). We see this in reasoning capabilities too: After o1&#8217;s breakthrough in mathematical reasoning, many dismissed Deepseek's R1 as a mere copy. But Deepseek's earlier papers reveal a different story: the Chinese AI lab had developed expertise in <a href="https://arxiv.org/pdf/2405.14333">formal verification</a> and <a href="https://arxiv.org/pdf/2408.08152">Monte-Carlo tree search</a> - key techniques for training reasoning systems - before O1 was published, perhaps indicating a pattern in how reasoning capabilities emerge.</p><p>Perhaps most tellingly, while many of OpenAI's best researchers have <a href="https://www.businessinsider.com/openai-leaders-who-left-since-2023-sam-altman-leadership-struggle-2024-9">departed</a> (Ilya Sutskever, Mira Murati, John Schulman, Alec Radford, Barret Zoph, Bob McGrew, Jan Lieke), it hasn&#8217;t affected the company&#8217;s valuation much&#8212;raising <a href="https://www.reuters.com/technology/artificial-intelligence/openai-closes-66-billion-funding-haul-valuation-157-billion-with-investment-2024-10-02/">$6.6 billion at a $157 billion post-money valuation</a> suggests we're far from the regime where breakthrough insights from star researchers are the limiting factor.</p><p>Second, it helps one to <strong>ignore weird forms of brittleness in the models, which aren&#8217;t going to matter in the end. </strong>Mainstream discourse about AI seems bizarrely eager to coalesce around the weaknesses of the models: &#8220;LLMs are stochastic parrots&#8221;, &#8220;<a href="https://www.theguardian.com/technology/article/2024/aug/06/ai-llms">the &#8216;reversal curse&#8217; meant AI was doomed to fail</a>&#8221;, &#8220;the models can&#8217;t count the number of letters in a word&#8221;, the models struggle when they are asked to perform tasks &#8216;backwards&#8217;; the models can&#8217;t do simple visual reasoning puzzles.</p><p>These debates misunderstand the &#8216;inside view&#8217; from the labs&#8212;their sole research focus is the next step on the critical path. Specifically, in the last 3 years, the returns to scaling pre-training have been so high that <strong>it was nigh-on unjustifiable to dedicate researcher time and compute to anything that wasn&#8217;t a) making this scale more, or b) figuring out the &#8216;big ideas&#8217; afterwards</strong>. Do you really think that AI labs&#8212;who have assembled a density of talent in the core research teams, not seen since the Manhattan Project&#8212;couldn&#8217;t solve the reversal curse if they needed to? The answer is that distracting The Eye of Sauron from the most important problems simply isn&#8217;t worth it.</p><p>The inverse of this, of course, is that it makes salient which challenges are worth paying attention to &#8211; Ilya Sutskever, a leading researcher who is paying attention to the critical path, notes that <a href="https://www.theverge.com/2024/12/13/24320811/what-ilya-sutskever-sees-openai-model-data-training">&#8220;pre-training as we know it will unquestionably end&#8221;</a>, and, as a result, we'll need to jump to the next paradigm.</p><h3>Concept, Scale, Apply: The Simplest Story of AI R&amp;D</h3><p>In the broadest of strokes, there are three stages needed to make progress on each of the five levels to AGI:</p><ol><li><p>Concept: Proving a novel idea works at all, even crudely</p></li><li><p>Scale: Engineering a proven concept into a deployment-ready system</p></li><li><p>Apply: Transforming the technology into systems that create real-world value and can be widely deployed</p></li></ol><p>We can track the progress on OpenAI&#8217;s levels to AGI so far:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!QpNO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64418595-7663-444e-8249-0717584a653d_1240x626.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!QpNO!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64418595-7663-444e-8249-0717584a653d_1240x626.png 424w, https://substackcdn.com/image/fetch/$s_!QpNO!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64418595-7663-444e-8249-0717584a653d_1240x626.png 848w, https://substackcdn.com/image/fetch/$s_!QpNO!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64418595-7663-444e-8249-0717584a653d_1240x626.png 1272w, https://substackcdn.com/image/fetch/$s_!QpNO!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64418595-7663-444e-8249-0717584a653d_1240x626.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!QpNO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64418595-7663-444e-8249-0717584a653d_1240x626.png" width="1240" height="626" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/64418595-7663-444e-8249-0717584a653d_1240x626.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:626,&quot;width&quot;:1240,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:130643,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!QpNO!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64418595-7663-444e-8249-0717584a653d_1240x626.png 424w, https://substackcdn.com/image/fetch/$s_!QpNO!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64418595-7663-444e-8249-0717584a653d_1240x626.png 848w, https://substackcdn.com/image/fetch/$s_!QpNO!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64418595-7663-444e-8249-0717584a653d_1240x626.png 1272w, https://substackcdn.com/image/fetch/$s_!QpNO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64418595-7663-444e-8249-0717584a653d_1240x626.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!5zR0!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43d8a670-e24c-4058-8c3b-cdb974441371_1240x740.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!5zR0!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43d8a670-e24c-4058-8c3b-cdb974441371_1240x740.png 424w, https://substackcdn.com/image/fetch/$s_!5zR0!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43d8a670-e24c-4058-8c3b-cdb974441371_1240x740.png 848w, https://substackcdn.com/image/fetch/$s_!5zR0!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43d8a670-e24c-4058-8c3b-cdb974441371_1240x740.png 1272w, https://substackcdn.com/image/fetch/$s_!5zR0!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43d8a670-e24c-4058-8c3b-cdb974441371_1240x740.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!5zR0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43d8a670-e24c-4058-8c3b-cdb974441371_1240x740.png" width="1240" height="740" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/43d8a670-e24c-4058-8c3b-cdb974441371_1240x740.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:740,&quot;width&quot;:1240,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:162062,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!5zR0!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43d8a670-e24c-4058-8c3b-cdb974441371_1240x740.png 424w, https://substackcdn.com/image/fetch/$s_!5zR0!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43d8a670-e24c-4058-8c3b-cdb974441371_1240x740.png 848w, https://substackcdn.com/image/fetch/$s_!5zR0!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43d8a670-e24c-4058-8c3b-cdb974441371_1240x740.png 1272w, https://substackcdn.com/image/fetch/$s_!5zR0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43d8a670-e24c-4058-8c3b-cdb974441371_1240x740.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Sometimes the "Concept" stage splits into two phases: first proving a raw concept works at all (Attention, CoT), then developing it into something that advances the critical path to AGI (GPT-2, o1). We can see this pattern emerging with AI Agency.</p><div><hr></div><h2>Agency</h2><p>The starting shot of this era may have begun with Anthropic's <a href="https://www.anthropic.com/news/3-5-models-and-computer-use">computer use</a>. As Dario Amodei <a href="https://lexfridman.com/dario-amodei-transcript/#chapter11_computer_use">explains</a> in a recent interview, they jury-rigged computer control by training Claude to analyse screenshots and output click locations and keyboard commands, which could be chained together in a loop (show image, get click location, execute, repeat) to enable basic computer interaction across operating systems.</p><p>It's remarkable that this works at all &#8212; and to be clear, it barely does &#8212; because these models were trained for conversation. They lack direct understanding of computer interfaces, struggle with persistent memory across screenshots, and must awkwardly communicate every action in English (&#8220;click at coordinates 342, 156&#8221;) rather than having native computer control. They&#8217;re limited by their context windows, making it hard to handle complex, long-running tasks. This is a scrappily cobbled together agent with existing systems &#8211; not one built from the ground up.</p><p>But that&#8217;s next.</p><p>This proof of concept showed how agency can be broken into building blocks: goal consistency&#8212; can the agent maintain its objective over time; planning&#8212;can the model break down tasks into smaller steps; memory&#8212;can the model retain context across long sequences; and tools&#8212; can the model interact with a computer and APIs?</p><p>The hardest challenge is goal consistency &#8211; how do we get the models to maintain their objective?</p><h3>Goal Consistency: Teaching agents to stay focused</h3><p>Chatbots have already learned a basic form of goal consistency: maintaining their role as helpful, honest, and harmless assistants throughout a conversation. This ability emerged from training on examples of good behavior (observational learning) and receiving feedback from both human and AI raters (interactive learning). These same principles could extend to maintaining focus on &#8220;long-horizon tasks&#8221; &#8211; where the agent needs to stay aligned with its objectives over extended periods and multiple interactions.</p><h3>Learning from examples</h3><p>AI agents start like inexperienced employees &#8212; they need to first learn what basic tasks look like before they can work independently. <a href="https://open.substack.com/pub/inferencemagazine/p/on-o1">Observational learning</a> provides this foundation by having models observe and copy human demonstrations, like watching recordings of people using computers to complete tasks. This teaches the model valid actions (clicking, typing, navigating interfaces) and basic workflows, just as a new hire might shadow a senior employee to learn the basics of their role.</p><p>The internet does not have enough data where people complete a diverse range of tasks on their computer. This will require deliberate data collection and construction. Companies can (and do) hire knowledge workers to record their work as training data, which allows models to learn both the specific tasks being completed and the broader patterns of how humans approach and execute complex workflows. For the less scrupulous, there are many options available. Companies could collect data directly from users, either as a condition of use for their operating system, or without their knowledge (previously, Microsoft&#8217;s <a href="https://www.theverge.com/2024/11/22/24302947/microsoft-recall-windows-insider-testing-dev-channel-click-to-do">AI-powered Recall feature</a> took screenshots at regular intervals). Another rich source of training data could come from the vast library of programming tutorials and screen recordings on YouTube, using AI to convert these visual demonstrations into structured sequences of computer actions.</p><p>But observational learning alone is insufficient for building capable agents. A model trained only on imitation can repeat patterns it has seen but lacks understanding of goals and struggles with novel situations. Most critically, pure imitation provides no feedback loop - the model has no way to know if its actions succeeded or failed.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> That's why observational learning serves as just the first step, followed by interactive approaches where agents can receive feedback on their actions and learn to achieve objectives rather than simply mimic behaviors.</p><h3>Interactive Learning: Human, AI and environmental feedback</h3><p>Just as a new hire moves from watching training videos to working with a mentor, AI agents progress from imitation to receiving direct feedback. This feedback &#8212; &#8220;reward signals&#8221; from the environment or human/AI raters &#8212; acts as a carrot and stick, teaching agents which actions help achieve their goals and which don't. While imitation provides the basic playbook, this interactive feedback loop is essential for agents to learn how to stay focused and adapt their approach as tasks evolve.</p><p>Inspired by the &#8220;think step by step&#8221; approach that improved language model reasoning &#8212; where models explicitly break down their thinking into smaller logical steps - we can have agents decompose complex tasks into verifiable subtasks. Think of a task like &#8220;set up a new web app&#8221; &#8212; this can be broken into concrete steps: creating directories, initialising repositories, installing dependencies, writing server code, and deploying. At each step, a verification system checks if that specific subtask was completed correctly: did the directory get created with proper permissions? Did the git repo initialize? Did dependencies install without errors?</p><p>This granular feedback at each step provides much clearer learning signals than only evaluating the final outcome.</p><p>In general, reward signals could come from both monitoring the environment and external feedback. The system can monitor basic system state - checking if files exist, if programs are running, and if there are any error messages. Beyond these basics, task-specific verification is possible through automated testing frameworks checking if code works, static analysis tools evaluating code quality, and performance metrics tracking factors like load times and memory usage.</p><p>This internal verification could be supplemented with human feedback. For each task, the system generates multiple different trajectories - different attempts at accomplishing the goal. Human raters can pick which attempts they prefer and give more fine-grained feedback to the models. They consider nuances about how the task was executed, not just whether it was complete: how clean was the code? How efficient was the solution? Was the reasoning logical and simple to follow? And this approach isn't limited to coding &#8212; you could imagine designing systems that let users rate how well an agent handled their email or organized their files at each step.</p><p>While AI systems can do some of this trajectory labelling too, human raters are (at least currently!) able to provide feedback on higher-level qualities and catch the subtlest errors or misaligned behaviors.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a></p><p>These verification and feedback systems, while still in early stages, suggest a path to robust goal-following.</p><h3>Memory: Remembering in &#8216;neuralese&#8217;</h3><p>Current language models have to process everything through their context window or on an English language scratchpad, which limits their &#8220;working memory.&#8221; Yet, similarly, the main ideas are in place. In a recent paper, Meta researchers <a href="https://arxiv.org/pdf/2412.06769">demonstrated</a> how agents could maintain ongoing &#8220;thoughts&#8221; and &#8220;memories&#8221; in a compressed neural form (sometimes called &#8220;neuralese&#8221;). Instead of English-language tokens, it can represent objectives and information in an efficient way native to its own processing. Just as Claude&#8217;s computer use demonstrated a proof-of-concept for AI agency, scaling neuralese could enable efficient memory for future AI agents.</p><h3>Planning: Scale what already works</h3><p>Similar to goal consistency, we get some amount of planning capability &#8220;for free&#8221; from existing models. We can improve this through learning from expert examples too&#8212;imagine hiring McKinsey consultants to break down projects into actionable steps, or implicitly learning from software engineers. We might even get this capability automatically as a byproduct of teaching models to maintain consistent goals over long horizons&#8212;teaching agents to reliably pursue objectives might naturally improve their ability to break down and organize complex tasks.</p><h3>Tool Use: Building native AI interfaces</h3><p>Current AI-computer interfaces are inefficient &#8211; language models that do use external tools, like ChatGPT with plugins, or Claude Computer Use spit out awkwardly formatted JSON files to call APIs or move mechanically around websites for humans.</p><p>The solution is fundamentally an engineering challenge. One perhaps overkill approach would be to encode computer actions directly into the model&#8217;s input/output space - rather than inputting and outputting English, the model could output neural patterns that directly map to system operations. But will probably not be necessary &#8211; there is lots of low-hanging fruit. We&#8217;re still using interfaces designed for humans, instead of the rigid &#8220;ask and respond&#8221; format, we could develop fluid protocols where models maintain continuous connections with tools, cache key information and share information efficiently.</p><h3>Putting It Together: Engineering Agents</h3><p>Somewhere in a data center in Arizona, thousands of AI agents are probably humming away on virtual machines, working through tasks, receiving feedback on their trajectories, and, increasingly, getting better and better. And that&#8217;s not to underrate the fiendish engineering challenge. But the main ideas are in place.</p><p>We can point to the key components: reinforcement learning for goal consistency, memory architectures that maintain state without token bloat, planning systems that decompose tasks effectively, and tool use interfaces designed natively for AI. None of these requires fundamental breakthroughs - we understand what needs to be built. The next phase is engineering these components to production scale - creating memory systems that work across hour-long tasks and millions of tokens, planning systems that break down complex tasks reliably, and tool interfaces or operating systems optimized for AI interaction.</p><p>And such agents, with goal-coherence, memory, planning and tool-use, is the foundation for building AI&#8217;s that can make real discoveries.</p><div><hr></div><h2>Invention</h2><p>To build AGI, AI systems must have some ability to make &#8220;conceptual leaps.&#8221; I like to conceptualise &#8220;invention&#8221; as the ability to spot useful similarities between distant ideas from just a few examples.</p><p>Creativity is just connecting seemingly disparate ideas and having better taste in which long distance connections might be fruitful.</p><p>Consider how a physicist solves a new problem: they might see just a few examples and recognise, &#8220;ah, this behaves like that system I studied last year.&#8221; They don't need thousands of examples - they can interpolate from a sparse set of observations to grasp the underlying pattern. The more expertise they develop, the better they become at making these leaps with less data.</p><p>But this is difficult to get from imitation alone.</p><h3>The Generator-Verifier Loop: Verifying which ideas work</h3><p>For Einstein-level conceptual leaps, raw generation capability isn&#8217;t enough &#8211; current models can already generate generate endless variations on existing ideas. What matters is developing the <em>meta-cognitive </em>ability of recognising which creative leaps are actually valuable.</p><p>The key idea is to combine powerful generators with ground-truth verifiers. The generator-verifier loop provides rapid, reliable feedback. Take mathematics: when a model proposes a novel proof, automated theorem provers can immediately verify if it works. This instant feedback helps the model learn which intuitive leaps are actually fruitful.</p><p>And, like with &#8220;goal consistency&#8221; in &#8220;Agency&#8221;, we can give this feedback step-by-step, rather than a single time at the end. A 2023 paper from OpenAI, &#8220;<a href="https://arxiv.org/abs/2305.20050">Let&#8217;s Verify Step by Step</a>&#8221;, shows that providing feedback on each step of a model's reasoning process, rather than just the final answer, dramatically improves performance on mathematical problems. This makes sense &#8211; you&#8217;ll learn more from a tutor who checks each step of your work, rather than just a right or wrong.</p><p>For this to work, verification has to be easier than generation.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> In some domains like math and coding, this is intuitive &#8211; where we can check if theorems are valid, or code compiles. But verification doesn&#8217;t have to come from external systems. AI models can be their own verifiers.</p><p>We already <a href="https://openai.com/index/deliberative-alignment/">see</a> &#8220;verifier&#8221; language models that check the reasoning of a &#8220;generator&#8217;s&#8221; reasoning steps. This AI-based verification unlocks new domains: a model trained on scientific papers could learn to evaluate whether a proposed hypothesis is consistent with known evidence, or train itself on journal reviewers comments.</p><p>This further implies that as models get better at verification in one domain, they can help train better generators in related domains. The result is a kind of bootstrapping process. Each advance in verification enables better training of creative leaps, which in turn enables more sophisticated verification. Once we solve the core engineering challenges of fast, reliable verification, we could see rapid progress in models' ability to make genuine discoveries.</p><h3>Synthetic Data: Using AI to generate training data</h3><p>This verification process doesn't need to happen within a single model or during inference. As we approach the limits of available internet data for pretraining, the future of scaling likely lies in synthetic data generation - using expensive but capable reasoning models to generate high-quality training examples that simpler models can learn from. DeepSeek demonstrated this with r1 and V3: rather than having V3 develop reasoning capabilities from scratch during inference, they used r1&#8217;s strong reasoning abilities to generate verified examples for V3's training (similarly, o1 was rumored to be primarily a synthetic data generation model).</p><p>This trades expensive inference-time compute for one-time pretraining compute - once a model learns these verified patterns during training, it can apply them quickly during inference. It's a way to front-load the exploration and verification cost rather than paying it repeatedly at runtime.</p><h2>The Path Forward: The Automated Research Fleet</h2><p>The last stage of OpenAI&#8217;s five levels is to have AIs run organizations &#8211; this will require massive coordination. A fully automated AI organization will not have individual AI models in chat windows, or even a single AI agent operating alone on a virtual machine, but instead an automated research fleet: some proving theorems, others reviewing literature, generating hypotheses, running experiments, analyzing results, and developing new paradigms.</p><p>But the labs are confident. As Paul Graham <a href="https://paulgraham.com/avg.html">reminds us</a>, read the job listings. Three months ago, OpenAI began <a href="https://x.com/polynoamial/status/1836872735668195636">hiring</a> for a new multiagent research team. Just this week, they've done the same for a new robotics team &#8211; the next frontier after software singularity.</p><p><strong>We will build AGI before we agree on its definition.</strong> For whatever metrics we choose, whatever capabilities we demand, the main ideas are already here &#8211; or soon will be, discovered by the automated research fleets that lie at the end of the critical path. The science is done. What remains is engineering.</p><div><hr></div><p><em>Acknowledgements</em>: <em>This piece benefited enormously from Jack Wiseman's extensive editing and substantial contributions throughout. Thanks also to Duncan McClements, Thomas Larsen, Eli Lifland, Daniel Kokotajlo, Somsubhro Bagchi, Niki Howe, Ariel Cheng, Nathaniel Li, Andrea Miotti, Samuel Ratnam, Sanskriti Shindadkar, Nat Kozak, Maximilian Nicholson, Xavi Costafreda-Fu, Philip Guo, Jacob Goldman-Wetzler, Jeremy Ornstein, Miles Kodama, Ollie Jaffe, Ananth Vivekanand, Saheb Gulati, Chris Pang, Xi Da, and Sudarshanagopal Kunnavakkam for their thoughtful feedback and suggestions. And thanks to the many others who provided valuable input and discussions.</em></p><p><em>All mistakes and oversights remain my own.</em></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>This does slightly simplify. Offline reinforcement learning learns from datasets of past experiences - each containing states, actions, and their resulting rewards. While this provides a feedback loop through historical data, it's limited to learning only from previously tried approaches. Without being able to actively test new strategies, agents struggle to develop truly robust goal-directed behavior that can handle novel situations. The main point still stands - agents need interactive learning to develop robust goal-directed behavior.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>And there are lots of engineering tricks and low-hanging fruit that boost performance. For example, for critical decisions requiring high confidence, rather than always using the same amount of compute, the system can choose to spend extra &#8220;inference-time compute&#8221; when faced with especially important goal-related choices. The system can delegate simpler tasks to cheaper models while using compute-intensive procedures - like running extensive simulations or generating and evaluating multiple solution attempts - for complex decisions. It could ask you, if it&#8217;s unsure. Or try to better elicit the goal that you <em>actually </em>want.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>And, in general it is! (In the most general case, P != NP, we hope &#8211; h/t Miles K). <br>In practice, though, there are exceptions. There are important exceptions to the "verification is easier than generation" principle. For instance, in DNA synthesis, verifying if a sequence works requires physically making it in a test tube - a process more costly/bottlenecked than computationally generating candidate sequences (though this could be solved by better simulation models). Similarly, in AI alignment, verifying if a model is genuinely safe and honest can be harder than training it to generate outputs, since you can't trust the model's own explanations (it might be deceptive), can't rely on other AI verifiers (they could also be deceptive), and humans may find it intractable to verify complex reasoning patterns.</p></div></div>]]></content:encoded></item><item><title><![CDATA[On o1]]></title><description><![CDATA[&#8220;If you look at the fractal structure of a snowflake, you might think that whoever made it did something impossibly intricate and difficult, but that building it piece by piece must somehow be possible, since someone did it.]]></description><link>https://inferencemagazine.substack.com/p/on-o1</link><guid isPermaLink="false">https://inferencemagazine.substack.com/p/on-o1</guid><dc:creator><![CDATA[Theo Horsley]]></dc:creator><pubDate>Fri, 17 Jan 2025 17:08:41 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!9o2z!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c61df20-b545-4c7b-9acb-75940496383f_1010x1010.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<blockquote><p>&#8220;If you look at the fractal structure of a snowflake, you might think that whoever made it did something impossibly intricate and difficult, but that building it piece by piece must somehow be possible, since someone did it. In fact, both statements are false: the way to make a snowflake is not to think in terms of its pieces but to know the laws of physics, have enough raw material and a large enough chamber, set the temperature, pressure, and humidity correctly, and wait for long enough. Furthermore, this is the only way to make snowflakes; trying to piece together a single one from little bits of ice is hopeless.&#8221; &#8212; <a href="https://youtu.be/CR45mBkSH7g?si=ClbwrWkRXjnO1Ll5&amp;t=4673">Dario Amodei</a></p></blockquote><div><hr></div><p>Much of contemporary AI research is, in some sense, downstream from the perspective that we should, where possible, 'let the compute figure it out'. By Moore's Law&#8212;or at least the folk version of it&#8212;the scale of the computational resources available are consistently and rapidly increasing. In the medium to long term, it's only the techniques which are able to most effectively leverage larger and larger quantities of compute that remain relevant. Hand-crafted methods tend to plateau and so are out-competed over time by general, flexible methods which, in figuring out how to do things for themselves, remain readily scalable. This perspective has been hard won, hence Rich Sutton dubbed it <a href="http://www.incompleteideas.net/IncIdeas/BitterLesson.html">The Bitter Lesson</a>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><p>This is a large part of why modern AI research is increasingly synonymous with the study of deep neural networks.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> Neural networks (or machine learning &#8216;models&#8217; as they are more often referred) are a kind of container which is able to hold a variety of possible '<a href="https://distill.pub/2020/circuits/zoom-in/">circuitry</a>'. It's this circuitry which determines how the model extracts and processes information from input to produce some output. If we have some metric which rates those outputs, we can, by a process of trial and improvement, search over the circuitry our model is able to hold so that, over time, our model 'learns' to perform well according to that metric. Importantly, the specific form that both our models and the particular process of trial and improvement<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> take allows them to be run and scaled up very effectively on modern computer hardware (GPUs and TPUs in particular), with improvements tending to come <a href="https://arxiv.org/abs/1502.01852v1">from</a> <a href="https://distill.pub/2017/momentum/">removing</a> <a href="https://arxiv.org/abs/1502.03167">bottlenecks</a> <a href="https://proceedings.mlr.press/v15/glorot11a/glorot11a.pdf">and</a> <a href="https://www.researchgate.net/publication/13853244_Long_Short-Term_Memory">improving</a> <a href="https://arxiv.org/abs/1706.03762">scalability</a>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a></p><p>Naturally, as the size of the container, the number and variety of examples, and the quantity of search performed scale with compute, the circuitry learnt, and thus our model's behaviour, can become increasingly sophisticated.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a> What our model learns will depend on what we ask it to do: what situations we put it in, what data we give it as input, which metric we use to grade its output, and so forth. Learning can be based on simply observing data or it can involve interaction, where past outputs can affect future inputs or feedback, depending on what we choose and what we want our model to learn.</p><div><hr></div><p>In the observational case, our hope is that our model can learn as much about the world as possible from the data it&#8217;s shown. In other words, we want our model to compress and internalise the structure and regularities it can find within the data. The nature of data the model is given (the format, the quantity, the quality) is thus the most important factor in determining what the model will learn. Depending on the data, there are two natural types of tasks we can give the model which will get it to compress: prediction (i.e. trying to guess what will come next) and reconstruction (i.e. trying to undo some process of noise or corruption). For both cases, <a href="https://blog.alexalemi.com/kl-is-all-you-need.html">there are natural metrics for our model to pursue</a>, with small changes depending on the exact set up we place our model in and what we&#8217;re asking our model to predict, or put differently, reconstruct.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a> Given the data and the task, as long as we then ensure that our model and training set-up are structured appropriately, then we should be able to readily scale it up: in the quantity of data that we feed it, in the various dimensions of our container, in the quantity of training we give it. As the model is scaled up, it will be able to pick up on more subtle patterns and structure within the data at a finer level of precision. If our data is sufficiently rich, then it will contain structure across many different resolutions, similar to how <a href="https://en.wikipedia.org/wiki/Coastline_paradox">coastlines remain rough</a> as you can resolve closer and closer. It&#8217;s for this reason, <a href="https://arxiv.org/abs/2102.06701">we</a> <a href="https://www.youtube.com/watch?v=MUvFuZpxLU8">think</a>, that we tend to find <a href="https://arxiv.org/abs/2001.08361">power</a> (or &#8216;scaling&#8217;) <a href="https://arxiv.org/abs/2203.15556">laws</a> between <a href="https://arxiv.org/abs/2010.14701">our natural metrics</a> and the aspects of increasing scale in the training of our model (i.e. model size, data quantity, training compute).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a></p><p>Prediction based learning on text, where the model is given some leading text then attempts to predict the continuation, has been particularly important for getting very powerful, general models. Text is the paradigm medium for efficiently representing world knowledge and reflecting on various aspects of thought. The vast range of subtlety and complexity in the human textual corpus has meant that the corresponding power laws have held over more than a dozen orders of magnitude. As our models moved to larger and larger scales, they internalise more and more, demonstrating deep knowledge and understanding of the world as well as some aspects of thought. Other aspects of thought, however, can be more difficult to internalise this way. In particular, abilities such as long term coherence, detecting and correcting errors, backtracking, and other &#8216;long horizon thinking&#8217; skills, have found difficulties arising from their effects being widely distributed over time and from there being few high quality demonstrations.</p><div><hr></div><p>Why do we care about this so much? Well, almost all real world applications involve extended tasks and so these issues have prevented these models from moving outside of areas where this doesn&#8217;t matter so much (chatbots, coding assistants, search) and into things like agents, which can go out and independently perform tasks in the real world. These &#8216;long horizon thinking&#8217; issues have been the primary bottleneck for getting agents to work well in the real world. Other research paths which were blocked by this such as the desire to gain the effects of a larger, more intelligent model by letting a weaker model think for longer. This is a particular instantiation of wanting to trade-off training compute and inference compute (i.e. compute used in running the model).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a> Also blocked were hopes of getting around the &#8216;data wall&#8217; - the issue that in pretraining, high quality data is increasingly scarce and expensive - by generating high quality synthetic data by thinking hard before creating the final result as well as providing higher quality feedback in other parts of training.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a></p><p>The primary reason that people have been so excited by o1 is that it was the first large-scale demonstration of a new (<a href="https://youtu.be/Nlkk3glap_U?si=fcliER025UYTZ6mX&amp;t=1125">though much anticipated</a>) technique which attempts to address precisely these concerns, potentially unblocking one of the major constraints of these models.</p><div><hr></div><p>In the interactive case, our hope is that our model can learn useful skills by experience from interacting with the world. Here there is no clear signal of good behaviour a priori as we had in the observational case. Under the formalism of reinforcement learning, we can separate out the problem by assuming that we have some reward signal which our model can receive after every output it makes and from which we can derive good behaviour by trying to maximise the total of our reward signal over time. Broadly, there are two kinds of learning we can do: &#8216;value learning&#8217; and &#8216;policy learning&#8217;.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-10" href="#footnote-10" target="_self">10</a> In value learning, the model learns the future reward it expects to collect given its current state<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-11" href="#footnote-11" target="_self">11</a> with the policy (the way that the actions actually taken in the world are selected) being learnt or derived from these estimates.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-12" href="#footnote-12" target="_self">12</a> In policy learning, the model learns to output actions directly by taking the total reward the model will receive <a href="https://proceedings.neurips.cc/paper_files/paper/1999/file/464d828b85b0bed98e80ade0a5c43b0f-Paper.pdf">to be our metric of pursuit</a>. This maximisation requires trialing our working policy at every step and so pure policy requires an overwhelming number of trials in the world. Thus value learning is added to make learning more efficient, with the particular art being to do <a href="https://arxiv.org/abs/2009.04416">each well</a> <a href="https://dl.acm.org/doi/abs/10.5555/3618408.3619934">without harming the other</a><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-13" href="#footnote-13" target="_self">13</a>, along with the ability to make use of trials collected from <a href="https://arxiv.org/abs/1502.05477">slightly</a> <a href="https://arxiv.org/abs/1707.06347">older</a> <a href="https://arxiv.org/pdf/2110.00641">versions</a> of your policy. When acting in more complex domains, <a href="https://arxiv.org/pdf/1808.00177">this latter</a> <a href="https://arxiv.org/abs/1912.06680">approach</a> <a href="https://deepmind.google/discover/blog/alphastar-mastering-the-real-time-strategy-game-starcraft-ii/">is</a> <a href="https://arxiv.org/abs/2303.08774">common</a>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-14" href="#footnote-14" target="_self">14</a></p><p>In some domains, such as games or question answering, identifying a suitable reward signal is straightforward (e.g. win / lose, video game score, correct / incorrect) but this is not true in general. For example, in using reinforcement learning to train for preferences where a separate &#8216;<a href="https://arxiv.org/pdf/1706.03741">reward model</a>&#8217; (itself trained to predict how a human or <a href="https://arxiv.org/pdf/2212.08073">another AI model</a> would rate some behaviour) generates a cheap reward signal for another model.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-15" href="#footnote-15" target="_self">15</a> Some reward signals (reward models in particular) run into issues of <a href="https://arxiv.org/pdf/2305.20050">robustness</a> where the model training against the signal exploits defects to get high reward without learning the intended behaviour. Rewards can also have varying degrees of sparsity (i.e. more or less frequent with the number of actions taken) with sparser rewards being more difficult to learn from, requiring good initial behaviour or a high degree of exploration.</p><div><hr></div><p>One way to attempt to address the issues around &#8216;long horizon thinking&#8217; skills<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-16" href="#footnote-16" target="_self">16</a>, is to set our model against some problem and allow it to think before answering. Using our reinforcement learning techniques, over time the model should learn how to leverage its ability to think to help it produce correct answers. Remembering the Bitter Lesson, we should also likely avoid interfering too much with how the model thinks and hoping that, by letting the model learn to think by itself, the kinds of skills we&#8217;re after will emerge<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-17" href="#footnote-17" target="_self">17</a>, though likely allowing the model to check intermediate runs (e.g. seeing if intermediate code runs). We should also try to make sure that what we&#8217;re doing is as scalable as possible. We want to pick problems which give some clear robust notion of completion and of these &#8216;verifiable&#8217; problems, ideally we want something where the verification is relatively quick and cheap so that they don&#8217;t become bottlenecks for our scaling.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-18" href="#footnote-18" target="_self">18</a></p><p>To ensure that our model has a signal it can learn from, we also need to ensure our problems are appropriately scoped, being neither too easy nor too difficult, so that our model has an incentive to improve its thinking. If the model knows little physics, we can't just give it a whole load of advanced magnetohydrodynamics questions to learn from, it'll never get any correct and won't know how to improve. So we're looking for a sweet spot where the problems we give the model to solve are neither too difficult nor too easy that they won't learn anything useful.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-19" href="#footnote-19" target="_self">19</a> While areas like mathematical proofs against checkers and coding against unit tests provide large classes of such problems, more broadly finding a large number of problems which satisfy these criteria can be difficult, so it&#8217;ll be important that our training method is very efficient in terms of the number of problems it needs to learn well.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-20" href="#footnote-20" target="_self">20</a></p><p>Ultimately, the value of this verifiable RL training depends on empirical degree of scalability, and how well the training generalises between domains (if our model has learnt to think through maths problems, will these skills transfer to thinking about other areas?) and timescales<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-21" href="#footnote-21" target="_self">21</a> (if our model has learnt how to perform well over an hour<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-22" href="#footnote-22" target="_self">22</a>, will our model also be able to perform well over a day? Over a week?). The degree of generalisation between domains, in particular, has special importance as it determines to what degree limited cheap verifiers are sufficient to get our skills more broadly or whether specific training will have to be done for each task or domain, but could increase with a sufficient diversity of tasks and scale.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-23" href="#footnote-23" target="_self">23</a></p><div><hr></div><p>While little is known about the details of these questions, we do know that in the case of whatever kind of training they&#8217;re using for o1, we know that the scalability is likely good<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-24" href="#footnote-24" target="_self">24</a> as within around 3 months of the announcement of o1, OpenAI announced their scaled up <a href="https://www.youtube.com/watch?v=SKBG1sqdyIU">o3</a> model. The consensus<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-25" href="#footnote-25" target="_self">25</a> seems to be that the primary improvement made here was a tremendous scale up of the RL training procedure (approaching the order of compute used for pretraining) and shows the kinds of results you would expect, with markedly strong performance of difficult mathematics and competitive coding.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-26" href="#footnote-26" target="_self">26</a></p><p>Zooming out, what does this mean in the big picture? Well, we might say that there is a third type of large scale progress to pay attention to in the frontier models, with advanced in these new verifiable RL (or &#8216;reasoning&#8217;) techniques now standing alongside the improvements in world knowledge and understanding from pretraining and the advancements in the domain these models can operate in from improvements like <a href="https://deepmind.google/technologies/gemini/">multimodality</a>, <a href="https://www.anthropic.com/news/3-5-models-and-computer-use">compute use</a>, and so on.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-27" href="#footnote-27" target="_self">27</a><br></p><p><em>Thanks to Toby Logan, Jack Wiseman, Seb Handley and Alicia Pollard for feedback on earlier drafts of this post.</em></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>It&#8217;s worth noting that beyond Moore&#8217;s Law, there&#8217;s the question of how much money you&#8217;re willing to spend on computational resources. As it has become clear that we may be close to the compute necessary for potentially transformative systems, this is the trend which has significantly accelerated since ~ 2018.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>One could argue that e.g. evolutionary or more explicit search techniques also follow our philosophy. In practice, however, these methods seem to scale less well (though, in some cases, they can act as valuable <a href="https://gwern.net/backstop">backstops</a> to other techniques) and thus we &#8216;let the compute figure out&#8217; which technique to use.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p><a href="https://pytorch.org/docs/stable/generated/torch.optim.SGD.html#torch.optim.SGD">Variations of</a> <a href="https://pytorch.org/docs/stable/generated/torch.optim.RMSprop.html#torch.optim.RMSprop">stochastic</a> <a href="https://pytorch.org/docs/stable/generated/torch.optim.Adam.html#torch.optim.Adam">gradient</a> <a href="https://pytorch.org/docs/stable/generated/torch.optim.AdamW.html">descent</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>Often improvements to efficiency running the model and unblocking the learning inside the models goes hand in hand. The <a href="https://arxiv.org/abs/1706.03762">transformer</a> is an important example, where the primary improvement arose from efficient use of parallelism on both sides.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>i) It&#8217;s not just that behaviour becomes more sophisticated, training itself often becomes more <a href="https://arxiv.org/abs/1912.06680">stable</a> and <a href="https://arxiv.org/pdf/1809.11096">robust</a> at larger scales.<br>ii) Such scaling does also involve changing other variables in the training procedure, though some of the <a href="https://arxiv.org/pdf/2001.08361">necessary changes</a> <a href="https://arxiv.org/abs/2203.03466">can be predicted</a> ahead of time.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>e.g. for prediction, are we asking the model to output a full distribution over outcomes or estimation of single value? Or for reconstruction, do we want the model to do a single reconstruction (as in a <a href="https://arxiv.org/abs/1312.6114">VAE</a>) or is it iterative (as in <a href="https://lilianweng.github.io/posts/2021-07-11-diffusion-models/">diffusion</a>)?</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>Though in other regimes, the scaling is less limited by resolution, as by noise and variance leading to a different scaling.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>For o1 in particular, the trade-off on AIME seemed to be roughly that a 10x scale up in inference compute allows you to train with about 20-ish? x less RL compute for constant performance (at least from poorly eyeballing the charts <a href="https://openai.com/index/learning-to-reason-with-llms/">here</a>, someone with a ruler should check me) (in comparison to AlphaGo where 10x more inference lets you get away with about 7x less training <a href="https://arxiv.org/pdf/2104.03113">compute</a>). That said, where this trade-off lands remains unclear (e.g. see <a href="https://www.lesswrong.com/posts/HiTjDZyWdLEGCDzqu/implications-of-the-inference-scaling-paradigm-for-ai-safety#MPNF8uSsi9mvZLxqz">this Gwern comment</a>).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p>This latter seems to have been the particular focus of <a href="https://youtu.be/OoL8K_AFqkw?si=csm3DcQJl69JFoLw&amp;t=585">Ilya Sutskever</a> and <a href="https://www.lesswrong.com/posts/HiTjDZyWdLEGCDzqu/implications-of-the-inference-scaling-paradigm-for-ai-safety#MPNF8uSsi9mvZLxqz">Gwern</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-10" href="#footnote-anchor-10" class="footnote-number" contenteditable="false" target="_self">10</a><div class="footnote-content"><p>I guess, one could imagine a technique purely based on a search atop a learnt reward model, however <a href="https://arxiv.org/pdf/1911.08265">in</a> <a href="https://arxiv.org/pdf/2301.04104">practice</a> policy and value learning is required to make this approach work well.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-11" href="#footnote-anchor-11" class="footnote-number" contenteditable="false" target="_self">11</a><div class="footnote-content"><p>As well as some initial action, in the case of <a href="https://www.nature.com/articles/nature14236">Q learning</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-12" href="#footnote-anchor-12" class="footnote-number" contenteditable="false" target="_self">12</a><div class="footnote-content"><p>i) This being a major source of instability in the pure value learning case.<br>ii) Value learning tends to show large benefits from frequent reuse of past samples (as long as those samples are sufficiently diverse), thus why value learning tends to use (often quite large) replay buffers which store past experiences.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-13" href="#footnote-anchor-13" class="footnote-number" contenteditable="false" target="_self">13</a><div class="footnote-content"><p>This comes from the fact that it can often be beneficial to have a single network which produces both policies and values, as they can both benefit from shared circuitry. However, as the policy and the value have different training profiles (e.g. very different <a href="https://arxiv.org/abs/1812.06162">critical batch sizes</a>) it is best to train the policies and values alternately in phases whilst ensuring that one isn&#8217;t too affected by the training of the other.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-14" href="#footnote-anchor-14" class="footnote-number" contenteditable="false" target="_self">14</a><div class="footnote-content"><p>i) As a passing note, it&#8217;s interesting to see that many of these large scale uses of these simple algorithms (like <a href="https://arxiv.org/abs/1707.06347">PPO</a> as well as something like pretraining if we broaden our scope) seem to be the main historical results published by OpenAI. We also note here that OpenAIs Chief Scientist Jakub Pachocki was the lead on both <a href="https://arxiv.org/abs/1912.06680">OpenAI Five</a> and <a href="https://arxiv.org/pdf/2303.08774">GPT-4</a> (though was not involved in the preceding work) as well as one of the leads on <a href="https://cdn.openai.com/o1-system-card-20241205.pdf">o1</a>. Increasingly, the focus has shifted to the kind of research required to execute scaling.<br><br>ii) These algorithms can often transfer quite well to the setting of having <a href="https://arxiv.org/abs/2103.01955">multiple agents</a> learning to interact with one another (<a href="https://www.youtube.com/watch?v=06VsbwJkrIo">though not always</a>).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-15" href="#footnote-anchor-15" class="footnote-number" contenteditable="false" target="_self">15</a><div class="footnote-content"><p>Preference based reinforcement learning is commonly used atop large models which have been &#8216;pretrained&#8217; on text prediction. As the pretrained model can generate text similar to what it&#8217;s seen during training, this kind of training tends to be thought of as eliciting particular skills, personalities and behaviours which are latent in the model.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-16" href="#footnote-anchor-16" class="footnote-number" contenteditable="false" target="_self">16</a><div class="footnote-content"><p>There are other approaches than the one described (e.g. the one used <a href="https://arxiv.org/pdf/2501.04519">here</a>), though I try to stick closer to what OpenAI have described and some guess work, though not too much is left to the imagination. I think that the biggest open question is what they&#8217;re doing with rewards. It could be a mix of end outcomes and <a href="https://arxiv.org/abs/2305.20050">process reward models</a> (i.e. reward models which inspect and give rewards based on the details of the model&#8217;s thought process), though I&#8217;m unsure. At least the <a href="https://openai.com/index/learning-to-reason-with-llms/">publicly released thought processes</a> don&#8217;t show strong signs of a PRM.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-17" href="#footnote-anchor-17" class="footnote-number" contenteditable="false" target="_self">17</a><div class="footnote-content"><p>i) Other than our &#8216;long horizon thinking skills&#8217;, we might hope that letting the model think for itself, it will be able to get used to and develop its own thinking style, beyond what it&#8217;s learnt from pretraining. Maybe this could help with things like &#8216;taste&#8217; and other kinds of subtle ways the model learns from the experience of practicing thinking. Afterall, it&#8217;s certainly the case that some things can only be learnt by doing them.</p><p>It&#8217;s also worth noting that this kind of training is very general. Many agentic tasks are, in fact, verifiable (you can check if your holiday has been booked or not), though such verification may be very expensive so you may or may not want to train on it depending on other factors (in particular, how much you can get away with other kinds of (potentially cheaper) learning as well as the degree of generalisation within our RL training)<br><br>ii) And indeed, in o1, this is what we <a href="https://openai.com/index/learning-to-reason-with-llms/">see</a>!</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-18" href="#footnote-anchor-18" class="footnote-number" contenteditable="false" target="_self">18</a><div class="footnote-content"><p>It&#8217;s interesting to consider the ramifications this has on the hardware side. Verifiers like this tend to run best on CPUs which suggests that this kind of training favours chips which have a higher CPU / GPU ratio. For more on what this actually means for e.g lab competitiveness, I&#8217;d recommend reading <a href="https://semianalysis.com/2024/12/11/scaling-laws-o1-pro-architecture-reasoning-training-infrastructure-orion-and-claude-3-5-opus-failures/">this SemiAnalysis piece</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-19" href="#footnote-anchor-19" class="footnote-number" contenteditable="false" target="_self">19</a><div class="footnote-content"><p>In the case of games, these sorts of reasons are a large part of why self-play is important, you always have a competitor at a similar skill level to yourself. Of course, then you need to do other things to ensure you&#8217;re learning against a sufficiently diverse set of adversaries and to encourage exploration of new strategies.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-20" href="#footnote-anchor-20" class="footnote-number" contenteditable="false" target="_self">20</a><div class="footnote-content"><p>Though not necessarily in terms of the attempts we make on our problems. Indeed, we should expect something like millions of trials as we scale up there. There are other kinds of RL techniques which are much more sample efficient in that regard but they tend to either use much more value based methods (which typically aren&#8217;t much used in training over pretrained language models, with more of the policy based methods being favoured instead (though I expect you&#8217;d want to use some policy based method with greatly improved value learning to help deal with sparse reward issues and there you&#8217;re likely to take inspiration from some of the sample efficient value based approaches)), or use world models (which aren&#8217;t really applicable here because &#8216;thinking in your head / in simulation&#8217; and &#8216;thinking aloud / in actuality&#8217; cost exactly the same for these models).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-21" href="#footnote-anchor-21" class="footnote-number" contenteditable="false" target="_self">21</a><div class="footnote-content"><p>I suspect that this is relatively poor, with the models really only being able to think on the same order of &#8216;time&#8217; (see next fn) as the most it&#8217;s been trained on. This however, is not too much of a concern as long as one can find problems which require such time for training (and if nothing else such problems can be added to the training data as they come up in practice). It only really matters if you&#8217;re after some small number of important problems (e.g. a Millenium Prize problem) which probably requires a lot of thinking, but you can&#8217;t justify the cost of the training. This doesn&#8217;t seem like too much of a big deal.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-22" href="#footnote-anchor-22" class="footnote-number" contenteditable="false" target="_self">22</a><div class="footnote-content"><p>This kind of time is a metaphorical comparison to a human. In models, the relevant measure is how much text (either in the model's thinking or in context) has the model seen or generated.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-23" href="#footnote-anchor-23" class="footnote-number" contenteditable="false" target="_self">23</a><div class="footnote-content"><p>Another thing to pay attention to is whether this kind of training generalises to training instead a single agent, systems of many copies of agents acting and thinking (potentially sharing residual streams or latent spaces) in a coordinated manner. This has the advantage of making potentially better use of inference compute in parallel instead of serially.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-24" href="#footnote-anchor-24" class="footnote-number" contenteditable="false" target="_self">24</a><div class="footnote-content"><p>Indeed, the <a href="https://openai.com/index/learning-to-reason-with-llms/">initial blog post</a> indicated strong example scaling laws for o1 on the AIME benchmark, though this is not necessarily the scaling relationship we most care about.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-25" href="#footnote-anchor-25" class="footnote-number" contenteditable="false" target="_self">25</a><div class="footnote-content"><p>Based on a mixture of staring at the few graphs shown and vibes, so take it or leave it as you please.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-26" href="#footnote-anchor-26" class="footnote-number" contenteditable="false" target="_self">26</a><div class="footnote-content"><p>e.g. <a href="https://www.youtube.com/watch?v=SKBG1sqdyIU">25.2 on EpochAI&#8217;s Research Math benchmark and 2727 on Codeforces Competition Code.</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-27" href="#footnote-anchor-27" class="footnote-number" contenteditable="false" target="_self">27</a><div class="footnote-content"><p>i) I first became aware of this framing from <a href="https://youtu.be/a0bEU83P8g8?si=3Z2Aam6nJtVoFYDk&amp;t=2872">Bob McGrew</a>, though for a similar framing see <a href="https://youtu.be/4a5lzYreMME?si=K2atZfe0rBxCk1he&amp;t=1109">this</a> from Jared Kaplan.<br><br>ii) Now that we don&#8217;t have a fundamental bottleneck from these long horizon skills, what are we supposed to say to ensure that we sound measured and reasonable? Well, at least in the short term, reliability will remain an issue. Very subtle and long horizon decisions that humans are able to make may be bottlenecks. &#8216;Taste&#8217; is one area where we might expect the generalisation across domains to be particularly poor. Potentially different forms of memory (surely there&#8217;s something you can do instead of just throwing the back of the KV cache away?). There does seem to be some inability in the models to refactor how they understand things? (Maybe there&#8217;s some weird layering or mixing of pre- and RL training which you can do here??) That said, it&#8217;s unclear how important or difficult to overcome these more speculative considerations may be.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[Getting AI datacentres in the UK]]></title><description><![CDATA[Why the UK needs to create Special Compute Zones; and how to do it.]]></description><link>https://inferencemagazine.substack.com/p/getting-ai-datacentres-in-the-uk</link><guid isPermaLink="false">https://inferencemagazine.substack.com/p/getting-ai-datacentres-in-the-uk</guid><dc:creator><![CDATA[Jack Wiseman]]></dc:creator><pubDate>Fri, 15 Nov 2024 01:44:24 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!K2M5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb36ac303-da1e-4ce3-bab7-1e44b8c1b546_3400x2400.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Here is some context on the current situation:</p><ul><li><p>Since 2012, the computational power used to train the largest AI models <a href="https://epoch.ai/data/large-scale-ai-models">has grown 100 million-fold</a>. This has become enormously energy intensive: the most recent model from Meta used an estimated <a href="https://epoch.ai/blog/can-ai-scaling-continue-through-2030">27 megawatts of power capacity</a>, which is approximately the same power required for <strong>88,000 UK households</strong>, approximately the same as York. If trends in computational power scale-up continue, <strong>by 2030</strong>, training the largest model would use <strong>2.15 times the UK&#8217;s entire energy generation capacity</strong>.&nbsp;</p></li></ul><ul><li><p>The <strong>AI developers are power constrained</strong>: Microsoft has committed to a 20-year power purchase agreement <a href="https://www.wsj.com/business/energy-oil/three-mile-island-nuclear-plant-reopen-7accde1f">to </a><strong><a href="https://www.wsj.com/business/energy-oil/three-mile-island-nuclear-plant-reopen-7accde1f">reopen Three Mile Island</a></strong>; <a href="https://www.ft.com/content/29eaf03f-4970-40da-ae7c-c8b3283069da">Google</a> and <a href="https://www.aboutamazon.com/news/sustainability/amazon-nuclear-small-modular-reactor-net-carbon-zero">Amazon</a> have entered partnerships with energy developers to <strong>construct new Small Modular Reactors</strong>; and Elon Musk&#8217;s x.ai has <strong>converted a factory in Memphis</strong> into a datacentre, and <a href="https://www.dwarkeshpatel.com/i/149705443/being-head-of-compute-at-an-ai-lab">put natural gas generators outside</a>. SemiAnalysis, a leading third party research group, estimates that AI datacentre power demand will grow by <strong><a href="https://semianalysis.com/2024/03/13/ai-datacenter-energy-dilemma-race/#the-real-ai-superpowers">40 gigawatts globally in the next two years</a></strong>, roughly 200 times the average power demand of Liverpool.</p></li></ul><ul><li><p>We investigated the feasibility of low-carbon power sources to power AI datacentres in the UK: either a combination of wind, solar, grid batteries, and natural gas backup;&nbsp; or using nuclear power. Our modelling found that the cost-minimising allocation of renewables would <strong>require over 200km<sup>2 </sup>of contiguous land area </strong><em><strong>per gigawatt</strong>, </em>which would need to be <strong>next to an LNG import terminal</strong>. Furthermore, this would have <strong>40% higher carbon</strong> emissions, and lead to <strong>27 times more &#8216;expected</strong> <strong>deaths</strong>&#8217; than using nuclear power, based on historical patterns.</p></li></ul><ul><li><p><em>However</em>, as things stand, <strong>no developer would choose to build an AI datacentre in the UK</strong>. The UK&#8217;s nuclear construction costs are <strong>4.5 times <a href="https://datawrapper.dwcdn.net/U9bFA/1/">higher</a></strong><a href="https://datawrapper.dwcdn.net/U9bFA/1/"> than South Korea</a>, and building reactors takes <strong>twice as long</strong>. Approving a new power plant has taken <strong>6 and 12 years</strong> in the last two instances, while approving a reactor with the same basic design in France took just <strong>3 years</strong>. If a datacentre operator wanted to connect to the grid instead of building new power, it would take <strong>up to 15 years</strong> to get a grid connection, and the <a href="https://www.gov.uk/government/statistical-data-sets/international-industrial-energy-prices">industrial electricity prices</a> are <strong>four times higher than in the US</strong> and <strong>45% higher than in France</strong>.</p></li></ul><ul><li><p>Despite these facts,<em> </em>the UK can make a series of reforms and, in just a few weeks of swift action, the UK can become <strong>the best place in the world to build AI datacentres</strong> and the nuclear power to support them. Special Compute Zones would create an alternative planning and regulatory approval process for nuclear power, AI datacentres, and the transmission and networking infrastructure to support them. If these reforms allowed UK projects to close the cost gap with South Korea by two thirds, nuclear power would become 37% cheaper than an equivalent blend of renewable power for datacentres. With high levels of speed and certainty during the approval process, the UK has the opportunity to catch the wave of AI infrastructure development.</p></li></ul><div><hr></div><h2>Navigating the report</h2><p></p><p>The <strong><a href="https://inferencemagazine.substack.com/i/151677344/overview">Overview</a></strong> can be read as a standalone. For more details&#8230;</p><ol><li><p><strong>&#8216;<a href="https://inferencemagazine.substack.com/i/151677344/ai-progress-is-very-quick">AI progress is very quick</a>&#8217;</strong>, explains what investors are buying into.</p></li><li><p><strong>&#8216;<a href="https://inferencemagazine.substack.com/i/151677344/how-ai-became-computationally-intensive">How AI became computationally intensive</a>&#8217;</strong>, offers a brief and simple technical introduction to why AI systems use so much computational power.</p></li><li><p><strong>&#8216;<a href="https://inferencemagazine.substack.com/i/151677344/further-progress-and-deployment-of-ai-systems-will-use-s-to-s-of-gigawatts">Further progress and deployment of AI systems will use 10&#8217;s to 100&#8217;s of gigawatts</a>&#8217;</strong> shows the historical trends in computational intensity and what forward-looking estimates suggest for the energy-intensity of AI.</p></li><li><p><strong>&#8216;<a href="https://inferencemagazine.substack.com/i/151677344/the-uk-would-power-ai-datacentres-using-nuclear-power-not-wind-and-solar">The UK would power AI datacentres using nuclear power, not wind and solar</a>&#8217;</strong> evaluates the suitability of low-carbon energy sources.</p></li><li><p><strong>&#8216;<a href="https://inferencemagazine.substack.com/i/151677344/going-without-ai-datacentres-would-be-a-mistake">Going without AI datacentres would be a mistake</a>&#8217;</strong> sets out the case for why the UK needs to have AI datacentres, and not depend on international markets.</p></li><li><p><strong>&#8216;<a href="https://inferencemagazine.substack.com/i/151677344/nuclear-power-is-slow-and-expensive-to-build-but-it-could-be-cheaper-and-faster">Nuclear power is slow and expensive to build, but it could be cheaper and faster</a>&#8217;,</strong> is a diagnosis of why UK nuclear construction costs are so high, and how this might be remedied.</p></li><li><p><strong>&#8216;<a href="https://inferencemagazine.substack.com/i/151677344/the-uk-should-create-special-compute-zones">The UK should create Special Compute Zones</a>&#8217;,</strong> suggests the implementation details for making the UK the best place in the world to build AI datacentres.&nbsp;</p></li></ol><p></p><p><em>We would like to extend thanks to John Myers, Robert Boswall, Nathaniel Read, Freddie Poser, Ben Southwood, Mustafa Latif-Aramesh, and Samuel Hughes; for their feedback and comments on this work. </em></p><p><em>We would also like to acknowledge the work of <a href="https://www.britainremade.co.uk/">Britain Remade</a>, in particular <a href="https://www.samdumitriu.com/">Sam Dumitriu</a>, whose high-quality analysis of nuclear power and infrastructure costs was instrumental in this proposal.</em></p><div><hr></div><h2>Overview</h2><p>Some kinds of growth are difficult to achieve. One way people create economic growth is by making a new discovery in a lab, doing the engineering work to turn this discovery into a product, and then running a business to distribute the product to the world. The academic literature suggests this kind of growth is getting harder over time&#8212;it requires adding <a href="https://web.stanford.edu/~chadj/IdeaPF.pdf">ever more researchers to maintain a consistent rate of breakthroughs</a>. This is what economists call &#8216;frontier&#8217; growth&#8212;this requires doing something nobody has done before. On the other hand, some other kinds of growth can be comparatively simple: an investor might come along, looking to repeat a technique proven to work elsewhere, and they'd like to build a factory of some sort to replicate it. Enabling this kind of &#8216;catch-up&#8217; growth is a choice.</p><p>The UK has been <em>without </em>growth for many years. Since 2007, per capita GDP has grown by just <a href="https://data.worldbank.org/indicator/NY.GDP.PCAP.PP.KD?locations=GB">0.35% per year</a> and total factor productivity growth <a href="https://ourworldindata.org/grapher/tfp-at-constant-national-prices-20111?time=2008..latest&amp;country=~GBR">has been flat</a>.</p><p>Sometimes, a new type of frontier growth will emerge: a combination of scientific breakthroughs will lead to the creation of a new <strong>&#8216;general purpose technology&#8217;</strong>. There have been three in modern history: the steam engine, electricity, and information technology (computers). These technologies will spread through many industries, and become a platform for the development of further technologies. Each general-purpose technology took decades to reach mass adoption, but each provided a huge opportunity for economic growth. The Industrial Revolutions were organised around a general-purpose technology: the First Industrial Revolution was steam power, the Second was electrification, and the Third was digitalisation.</p><p><strong>Such a breakthrough is happening today</strong>, in artificial intelligence (AI).</p><h4>AI is poised to do for cognitive labour what the steam engine did for physical labour</h4><p>One way to think about the impact of AI is using the steam engine as an analogy. In the First Industrial Revolution, there was a clear input and output relationship&#8212;pour in coal, and get out rotational power in a crankshaft, which could be used for all sorts of downstream tasks (trains, factories, <em>more </em>coal mining). What makes this so powerful is that it is <em>dependable</em>&#8212;every time one adds coal, they know what will happen; it is <em>general</em>&#8212;the rotational motion can be applied to any of these use cases; and it is <em>scalable&#8212;</em>so long as one can design a bigger steam engine, there is no limit to the amount of coal an engine can usefully convert into power.</p><p>A latent resource we have today&#8212;just like the British in the second half of the eighteenth century had a lot of latent energy in surface coal&#8212;is computational power in computer chips. Over the last 60 years, the number of transistors (think: computational power units) on a chip has doubled every<em> </em>two years, giving us an enormous supply of computational power. One intuition for what the field of AI is trying to answer is, &#8220;How can we exchange this computational power for useful information processing capabilities?&#8221;.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> This requires finding the system of pipes and valves&#8212;or in this case, the combination of architectures, algorithms, data, and the training procedure&#8212;to make the input-output relationship work.</p><p>Over the past 12 years, AI researchers have found the broad combination of variables which allow them to add more input. Today&#8217;s state-of-the-art systems score very highly on capability benchmarks. They <a href="https://openai.com/index/learning-to-reason-with-llms/">outperform </a><strong><a href="https://openai.com/index/learning-to-reason-with-llms/">PhD-level experts</a></strong> on a benchmark testing for scientific expertise; <strong>score in the 99th percentile</strong> on the US law school admissions test (LSAT)<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>; and are capable of winning a <strong><a href="https://deepmind.google/discover/blog/ai-solves-imo-problems-at-silver-medal-level">silver medal on the International Mathematics Olympiad</a></strong> exam (for the best sixth-form age students in the world). These capabilities have translated into <strong>economically useful performance</strong>: on <a href="https://www.swebench.com/">software engineering tasks</a>, in <a href="https://www.nber.org/papers/w31161">customer support</a>, and in <a href="https://aidantr.github.io/files/AI_innovation.pdf">materials discovery</a>.&nbsp;</p><p>AI systems will support economic growth in three main ways:</p><ol><li><p>AI will let us <strong>automate </strong><em><strong>existing </strong></em><strong>information processing tasks</strong></p></li><li><p>As the cost of information processing declines to zero, we will <strong>use this new abundance to create new products and services</strong>.&nbsp;</p></li><li><p>AI can help to <strong>accelerate research and development</strong> to drive frontier growth.</p></li></ol><p>What is important to understand about the input-output relationship of AI systems is that there are extremely<em> </em>dependable improvements <em>on the specific task</em> they have been trained to do. However, general purpose AI systems have been trained primarily to predict the next word in a sequence, and so it is <strong>genuinely uncertain whether improvements at text prediction will continue to transfer to useful capabilities</strong> in the system; and by extension, whether these will transfer to economically useful tasks. OpenAI&#8217;s GPT-3, the first system which most people had experience with through ChatGPT, had useful capabilities&#8212;it was good at writing sonnets&#8212;but it was <em>GPT-4</em> that could begin to do coding. So far these improvements have continued to transfer, which explains the impressive capabilities and economic interest in AI systems.</p><h4>The computational and energy-intensity of AI is only getting greater</h4><p>Because of this uncertain exchange rate between compute and &#8216;capabilities&#8217; broadly, AI developers want to <strong>continue increasing the computational inputs to create and run these systems, </strong><em><strong>and </strong></em><strong>they want to sell the current capabilities as broadly as possible</strong>. The challenge is that increasing computational power is enormously energy-intensive. AI systems use specialised chips, predominantly the Graphical Processing Unit (GPU) designed by NVIDIA. Each state-of-the art GPU uses an enormous amount of power: if used continuously, it <strong>will consume 44% more power than an average UK resident.</strong><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> To support this energy and computationally intensive process, AI chips are hosed in datacentres. These are specialised warehouses which are optimised to manage the cooling and energy requirements of the chips. From the outside, these are ordinary buildings on an industrial park.&nbsp;</p><p>The most recent AI system from Meta (called Llama 3.1) was trained <strong><a href="https://scontent-lhr8-1.xx.fbcdn.net/v/t39.2365-6/463020162_522238820565582_8192401983671993921_n.pdf?_nc_cat=108&amp;ccb=1-7&amp;_nc_sid=3c67a6&amp;_nc_ohc=6V_W4zoVlq0Q7kNvgHV6_Gk&amp;_nc_zt=14&amp;_nc_ht=scontent-lhr8-1.xx&amp;_nc_gid=ATFiT-sFUQpdhWuVQNOzvPZ&amp;oh=00_AYBeYjCOR6Cacf8E7_9-VHm43tEwXPvebkElCLjnnvHmUw&amp;oe=673BE619">using 16,000 active GPUs</a> in a single datacentre</strong>, which is <strong><a href="https://epoch.ai/blog/can-ai-scaling-continue-through-2030">estimated to have required 27 megawatts</a></strong> (MW) of installed power capacity. This is roughly equivalent to <strong>88,000 UK households</strong>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a> One useful intuition is that a single datacentre will use as much power as a small city.</p><p>As developers are betting on the input-output relationship holding, computational power is being scaled up at breakneck speed&#8212;the amount of computational power used in training the AI systems has grown <strong><a href="https://epoch.ai/blog/training-compute-of-frontier-ai-models-grows-by-4-5x-per-year">5 times over per year since 2017</a></strong>. Were the current trend to hold until 2030, training the largest AI system would require <strong>approximately 2.15 times the UK&#8217;s </strong><em><strong>entire </strong></em><strong>electricity generation</strong>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a> Meta&#8217;s release of Llama 3.1 in July 2024 might begin to look very small indeed! Running the systems, too, is set to become more energy intensive: new methods have been developed to input additional computational power whilst the model runs, to enhance capabilities further. Jensen Huang, the CEO of NVIDIA, expects this to grow by <strong>&#8220;<a href="https://x.com/RihardJarc/status/1845453408557289653">a billion times</a>&#8221;</strong>. More details on these trends are covered in the main body.</p><h4>There is a wave of capital investment to build 10s to 100s of gigawatts of datacentres and power for the next decade</h4><p>To continue increasing the input of computational power, the AI developers and the &#8216;cloud providers&#8217; who sell them datacentre capacity, are growing their operations as quickly as possible. SemiAnalysis <a href="https://semianalysis.com/2024/03/13/ai-datacenter-energy-dilemma-race/">estimates the power demand from AI datacentres</a> globally <strong>will grow 40 gigawatts (GW) by 2026, and in the US alone, by 47.8 GW for 2028, from just 8.5GW in 2024</strong>. To put this into perspective, one GW of continuous power demand is five times the average power demand of Liverpool.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a>The <strong>addition of AI datacentres globally in the next two years will use roughly 200 times the power demand of Liverpool</strong>.</p><p>This enormous surge in demand is constrained by energy generation. There is simply not enough power. As a result, AI developers are signing long-term power purchase agreements to bring nuclear power plants back online, and are signing development agreements with developers of &#8216;Small Modular Reactors&#8217; (SMRs). Perhaps most dramatically, Elon Musk&#8217;s company, x.ai, has even converted a factory in Memphis into a datacentre for 100,000 GPUs in just 19 days, and is <strong><a href="https://www.dwarkeshpatel.com/p/dylan-jon">using natural gas generators</a> outside</strong> to make up the power it needs.&nbsp;</p><h4>No developer would build an AI datacentre in the UK by choice</h4><p>As things stand, the coming wave of capital investment will bypass the UK. No AI developer or cloud provider would choose to build an AI datacentre with new power in the UK:</p><ul><li><p><strong>Planning permission for the datacentre would take too long</strong>. Until recently, datacentres went through Local Planning Authorities under the Town and Country Planning Act 1947, but the Labour Government is going to make them <a href="https://www.telegraph.co.uk/business/2024/06/09/labour-plots-to-build-data-centres-on-green-belt/">&#8216;Nationally Significant Infrastructure Projects&#8217;</a> which require a Development Consent Order from the Secretary of State, with the aim of accelerating the process. However, the <a href="https://www.samdumitriu.com/p/why-britain-struggles-to-build-infrastructure">average period of consideration</a> for a Development Consent Order in 2020 was <strong>22 months</strong>, and since the last election, ministers have <a href="https://www.thetimes.com/uk/politics/article/labour-ministers-delay-40-per-cent-of-infrastructure-projects-plfv9lbwj">delayed 40% of decisions</a> on Development Consent Order decisions, and so it is likely there would be further delays.</p></li><li><p>If a datacentre operator wanted to use grid power, it would take<strong> up to 15 <a href="https://www.nytimes.com/2024/01/02/business/uk-economy-growth.html?searchResultPosition=1">years</a> to get a grid connection</strong>, and even then the UK&#8217;s <a href="https://www.gov.uk/government/statistical-data-sets/international-industrial-energy-prices">industrial electricity prices</a> are<strong> four times higher</strong> than the US and <strong>45% higher</strong> than France.</p></li><li><p>If the datacentre operator wanted to procure their own nuclear power, <strong>it would take 6 to 12 years to get approval</strong>, and once they have approval, construction would take 12 to 15 years.</p></li></ul><p>The current pace of planning, regulatory approval, and construction is too slow to keep pace with the wave of investment.</p><h4>If the UK wants AI datacentres, nuclear power would be the safest, cleanest, least land-intensive, and could also be the cheapest.</h4><p>We investigated which energy generation method would be most suitable to power AI datacentres. We compared the feasibility of nuclear power, or a blend of wind, solar, grid battery storage, and natural gas backups.</p><p>We calculated the cost-minimising way to blend renewables with batteries and gas, to provide the permanent power supply datacentres require, and found that <em>per gigawatt </em>of firm power<em>, </em>it would require 8 GW of installed solar power, 0.37 GW of wind power, 12 GWh of battery backup, complete gas backup, and LNG import capacity for 451 million cubic metres of gas. This would become infeasible on multiple grounds:</p><ol><li><p>Land intensity&#8212;8 GW of solar panels and 0.37 GW of wind turbines would require <strong>160km<sup>2</sup> and 41km<sup>2</sup> respectively</strong>. (As a point of reference, Cardiff is 140km<sup>2</sup> and Reading is 40km<sup>2</sup>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a>) This <strong>scales very poorly</strong>. Some datacentre campuses are much larger than one gigawatt, for example, to power <a href="https://www.theinformation.com/articles/microsoft-and-openai-plot-100-billion-stargate-ai-supercomputer">Microsoft and OpenAI&#8217;s 5GW datacentre campus</a> in Virginia <strong>it would require over 1000km<sup>2</sup> of contiguous land.</strong></p></li><li><p>Emissions&#8212;because of the intermittency of wind and solar, natural gas would generate 28% of the power, which <strong>would produce 40% more carbon than equivalent nuclear capacity</strong>.</p></li><li><p>Safety&#8212;the air pollution from natural gas emissions would be <a href="https://ourworldindata.org/safest-sources-of-energy">many times more dangerous</a> than the risk of a nuclear accident, <strong>resulting in 27 times as many expected deaths</strong>.</p></li><li><p>Limited proximity to LNG import capacity. The UK has three LNG import terminals, one in Kent and two in South Wales. The natural gas generation terminal would need to be some distance from population centres to reduce air pollution, but also close enough to the import terminal that gas pipelines are not necessary. It is either necessary to find ways to build these generation facilities in Kent or South Wales <em>while being contiguous </em>with the solar and wind farms, or it might be necessary to build a new LNG import terminal, but neither approach scales well, and the latter would add substantially to costs.&nbsp;&nbsp;</p></li><li><p>Cost&#8212;we calculate a levelised cost of blending wind, solar, batteries, and natural gas at &#163;106/MWh, which is lower than the current CfD price for Hinkley Point C (&#163;143/MWh<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a>), though there are large opportunities for cost savings with nuclear power&#8212;South Korea builds at roughly 25% of the cost of the UK&#8212;and so if two thirds of the cost gap with South Korea could be bridged, the cost would be 37% cheaper than renewables. (In Texas, renewable energy at the same emissions intensity for &#163;74/MWh.)</p></li></ol><p>Building nuclear power plants in the UK has been slow and expensive, but it doesn&#8217;t have to be. Some relevant facts:</p><ul><li><p><strong>Two thirds of the cost of Hinkley Point C was interest</strong>&#8212;if you can bring down the cost of borrowing, this can cut the final cost of electricity in half.</p></li><li><p>South Korea builds <strong>8 to 12 copies of the same reactor design</strong>. This means they benefit from learning, both technically, and in terms of the regulation, <em>and</em> they have a consistent supply chain of components and of people with the skills to build a reactor. On the contrary, <strong>Hinkley Point C was a one-of-a-kind reactor</strong> which had <strong>7,000 design changes</strong> from the basic design already used in France and Finland, and was the first nuclear power built in the UK in 21 years.</p></li><li><p>Responsibility for nuclear power plant approval has been <strong>diffused between many actors who can say &#8216;no&#8217;</strong>, or who might add incremental delays and cost increases to new nuclear power plant construction, which amounts to a <em>de facto</em> &#8216;no&#8217;; but <strong>there</strong> <strong>is</strong> <strong>no</strong> <strong>positive force in the system</strong> who pushes for power plants to be built.</p></li></ul><p>Small Modular Reactors (SMRs) provide a big opportunity for the UK. Because most assembly happens in a factory, large productivity gains during manufacturing are possible, and on-site construction can take just months. Furthermore, SMRs are especially suited to AI datacentres because there can be a fleet powering a datacentre campus, and so when there is an outage, there is a diversified power supply.</p><p>Making UK nuclear power competitive is the only way the UK would attract AI datacentres: if a datacentre provider wanted to use a blend of renewables, they would be likely to be much better off in West Texas.</p><p>It is possible to bring the costs down&#8212;there is a lot of low-hanging fruit to be picked!</p><h4>The UK needs AI datacentres for economic security, growth in former industrial areas, and to seize the opportunity of future frontier growth</h4><p>At this point, a sceptic might ask whether it is worth trying at all. In general purpose technology revolutions, most of the gains come from <em>adopting </em>the new technology, and seem unlikely to accrue to those who host AI datacentres. Perhaps this wave of investment is going to happen, but can&#8217;t the UK just focus on &#8216;high-value&#8217; activities, like integrating AI systems and building AI applications.&nbsp;</p><p>We don&#8217;t think so&#8212;first and foremost, there&#8217;s no scarce resource being used up by permitting this growth, all the capital is from private investors, and the UK <em>needs </em>the growth. Most importantly, however, the UK needs the critical inputs for its future economic engine, the old economic dogma&#8212;that it is possible to sit atop the value chain and focus only on the &#8216;high value&#8217; activities&#8212;is incorrect. This causes the hollowing out of industries, and neglects the valuable &#8216;learning-by-doing&#8217; that allows us to make future growth. The UK without datacentres will lack the critical inputs for its future economic engine.</p><p>The Chancellor Rachel Reeves has frequently promoted a new doctrine of <a href="https://static1.squarespace.com/static/64f707cf512076037f612f60/t/6502d760c087cb1853b8f5c4/1694685033194/A+NEW+BUSINESS+MODEL+FOR+BRITAIN_0.pdf">&#8216;securonomics&#8217;</a>, of which the core tenets are prioritising <strong>economic security in an &#8216;age of uncertainty&#8217;,</strong> not <strong>depending on a narrow set of industries </strong>from London and the South East to drive growth for the whole country, and <strong>seizing the opportunities</strong> of a rapidly changing world.</p><p><strong>Hosting AI datacentres in the UK is central to all three tenets.</strong>&nbsp;</p><p>First, as AI systems become increasingly integrated into the economy, especially into the UK&#8217;s professional services export businesses, <strong>a large fraction of the UK&#8217;s capital stock will be created in, stored in, or run in AI datacentres</strong>. The UK will want these AI datacentres to be here, rather than overseas and connected through an undersea cable, to ensure it can protect these assets. Furthermore, as adoption is<em> </em>critical to capturing the gains, the UK needs to ensure it has computational power capacity it needs. As demand for computational power rises globally, it could be the case that UK businesses are unable to access this. The Microsoft CFO said on an <a href="https://www.microsoft.com/en-us/Investor/events/FY-2025/earnings-fy-2025-q1">earnings call two weeks ago</a> that revenue growth in their cloud business was 33% but, &#8220;[d]emand continues to be higher than our available capacity.&#8221;</p><p>Second, the <a href="https://www.gov.uk/government/publications/artificial-intelligence-sector-study-2023/artificial-intelligence-sector-study-2023">Government&#8217;s AI Sector Study</a> shows that 75% of UK AI companies are based in London, the South East, or East of England. This is to be expected: AI application developers will agglomerate around London because it has the best venture capital ecosystem and AI talent density outside San Francisco. However, as the Chancellor <a href="https://labour.org.uk/updates/press-releases/rachel-reeves-securonomics/">has said</a>, there has been, &#8220;[a] misconceived view held that a few dynamic cities and a few successful industries are all a nation needs to thrive&#8230;[t]he result was a paucity of ambition for too many places, the hollowing out of our industrial strength and a tragic waste of human potential across vast swathes our country.&#8221; It is now very rarely the case that growth can be so readily directed towards areas with a strong industrial past, but that is the opportunity of AI datacentres&#8212;it is possible to bring the Fourth Industrial Revolution to the rest of the UK as well, if the rules allow it.</p><p>Finally, given that AI systems are likely to play a critical role in research and development. The UK has world-leading science and technology clusters, whose work is likely to be transformed by AI systems. Running AI systems that support research and development&#8212;AI for science&#8212;<strong>will be critical to any frontier growth in the UK for decades to come</strong>, and it is not possible to depend upon international datacentre markets to supply the services which are so central to our future prosperity. They would need to be here.</p><p>The UK has already missed an opportunity for frontier growth&#8212;UK scientists like Geoffrey Hinton and Demis Hassabis were at the forefront of the AI revolution, which is now being commercialised by a small handful of US firms. The UK is about to pass up the opportunity for the easiest kind of growth&#8212;someone wants to build AI datacentres and power to support it. <em><strong>The money wants to flow</strong></em> but the revealed preference of our current regulatory and planning system is that it should not.&nbsp;</p><h4>The UK can become the world&#8217;s best place to build an AI datacentre and the power to support it</h4><p>This can all be fixed in as little as a few weeks. To do this, we propose creating &#8216;Special Compute Zones&#8217;&#8212;which provide an alternative planning and regulatory approval process that fixes the issues with the current approach. It would provide the <strong>certainty, speed, and hence, the opportunity to be cheap,</strong> which is currently lacking.</p><p>Developers could receive a <strong>single point of signoff</strong> for the power, datacentre, and transmission and networking infrastructure they require. Within the Zones, there would be <strong>&#8216;deemed consent&#8217;</strong>&#8212;meaning the default answer to construction is &#8216;yes&#8217;&#8212;and permission to construct would have to be challenged within three months of an application. Ordinarily, a planning decision will weigh the relative merits of each project; but within a &#8216;Special Compute Zone&#8217;, it would be decided that <em>by creating a zone, </em>this cost-benefit is pre-considered, and therefore approval would depend on a <strong>&#8216;condition-based approach&#8217;</strong>&#8212;if a developer can show that the project meets particular standards, it goes ahead. There is <strong>precedent for this kind of approach in the EU and Spain</strong>, where &#8216;Renewable Acceleration Areas&#8217; use condition-based approaches. We present more details on implementation below.</p><h4>Missing this wave of capital investment is like missing the railways</h4><p>Between 1841 and 1850, private investors ploughed a cumulative <strong><a href="https://x.com/michael_nielsen/status/1782129830932210166">40% of UK GDP into building the railways</a></strong>&#8212;imagine if instead our planning and regulatory regime had prevented this investment and the UK&#8217;s rapid economic growth! The UK continues to collect the dividend from this period of growth today, 170 years on.</p><p>Sometimes growth is difficult to come by, but in this case, growth is a choice: <strong>all we need to do is unhobble ourselves.</strong></p><p></p><div><hr></div><h2>1. AI progress is very quick</h2><h4>The goal of AI research is generally intelligent systems</h4><p>To begin, it is useful to clarify what AI research is aiming at, as people use many different terms. These include Artificial General Intelligence (AGI), Artificial Superintelligence (ASI), human-level AI, powerful AI, and transformative AI. The terms can be somewhat misleading&#8212;does &#8216;human-level AI&#8217; refer to AI systems which perform at the level of the <em>average </em>human, or the <em>smartest </em>human? Furthermore, the development of AI systems is &#8216;unbalanced&#8217;, in some ways current systems already surpass the smartest humans, but in other ways they fall far short.&nbsp;</p><p>Debates over these definitions can distract from focusing on the most important thing: very capable systems might be created before we have clarified whether &#8216;true&#8217; AGI requires, say, emotional intelligence. To avoid these pitfalls, three terms can be useful:</p><ul><li><p>A <strong>Drop-In Remote Worker</strong> refers to an AI system that can interact with a computer, pursue tasks for weeks-equivalent of human time, at the level of a graduate remote worker.</p></li><li><p>An <strong>Expert Scientist</strong> refers to an AI system that can perform scientific research, at the level of the world's best scientists across a variety of scientific domains.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a></p></li><li><p>A <strong>Superintelligence </strong>refers to an AI system that exceeds human intellectual capabilities across all relevant cognitive skills.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-10" href="#footnote-10" target="_self">10</a></p></li></ul><p>The explicit goal of the AI research labs is to create a software program that is an Expert Scientist. When DeepMind was founded in 2010, their mission <a href="https://www.theguardian.com/technology/2016/feb/16/demis-hassabis-artificial-intelligence-deepmind-alphago">was</a>: &#8220;To solve intelligence, and then use that to solve everything else.&#8221;</p><h4>AI researchers are making progress towards this goal</h4><p>Whether expert scientists are possible in the current technical paradigm is genuinely uncertain, but current state-of-the-art systems can do a lot:</p><ul><li><p><strong>Coding assistance. </strong>State-of-the-art AI systems are very effective at giving assistance to professional software engineers. <a href="https://ar5iv.labs.arxiv.org/html/2302.06590">This research</a> found that software engineers were 55.8% faster at completing a software-engineering task with assistance from an AI system, and the former head of self-driving at Tesla, Andrej Kaparthy, <a href="https://x.com/karpathy/status/1827143768459637073">wrote</a> that he, &#8220;basically can't imagine going back to &#8216;unassisted&#8217; coding at this point&#8221;.</p></li><li><p><strong>Scientific capabilities</strong>. State-of-the-art systems outperform experts with relevant PhDs on <a href="https://openai.com/index/learning-to-reason-with-llms/">GPQA diamond</a>, an evaluation which tests expertise in physics, chemistry, and biology<strong>.</strong></p></li><li><p><strong>Mathematical abilities. </strong>DeepMind&#8217;s AlphaProof was trained to solve problems from the International Mathematical Olympiad, and <a href="https://deepmind.google/discover/blog/ai-solves-imo-problems-at-silver-medal-level">achieved &#8216;silver medalist&#8217; performance&#8212;</a>roughly equivalent to scoring in the top 100 sixth form mathematicians in the world. OpenAI&#8217;s &#8216;o1&#8217; was not trained specifically to perform well at maths questions, but <a href="https://openai.com/index/learning-to-reason-with-llms/">scored 80%</a> on a US Mathematics Olympiad qualification exam, which is equivalent performance to the top 500 high school students in the USA.</p></li><li><p><strong>Agentic improvements</strong>. AI systems can complete small software engineering projects. <a href="https://www.swebench.com/">SWE-Bench Verified</a> is a benchmark which measures the ability of AI systems to perform real world software engineering tasks: OpenAI&#8217;s GPT-3.5, the state-of-the-art model in 2022 performed poorly, only able to complete 0.4% of the tasks. As of November 2024, Anthropic&#8217;s Claude 3.5 Sonnet was able to successfully complete 53% of tasks. For another example, OpenAI&#8217;s o1 <a href="https://cdn.openai.com/o1-system-card-20240917.pdf">successfully completed 100% of the problems</a> posed to interviewees for software engineering positions at OpenAI.</p></li></ul><h4>Where AI progress goes from here is very uncertain</h4><p>While the historic trajectory of AI progress is clearly very steep, it is difficult to know whether this implies that capability improvements will continue, and <em>exactly </em>what this might look like. Most importantly, we do not have a large suite of benchmarks for assessing the capabilities of the most advanced AI systems. Because the AI systems have improved so quickly, our evaluations &#8216;saturate&#8217;, meaning that all systems score indistinguishably high scores. For example, <a href="https://arxiv.org/pdf/2009.03300v3">MMLU</a> and <a href="https://arxiv.org/pdf/2103.03874v2">MATH</a> were benchmarks released in 2020 and 2021, and specifically designed to resist this kind of saturation. GPT-2 scored 32.4% on <a href="https://arxiv.org/pdf/2009.03300v3">MMLU</a> and 6.9% on <a href="https://arxiv.org/pdf/2103.03874v2">MATH</a> at their release, but in just three years, they approached saturation: <a href="https://cdn.openai.com/papers/gpt-4.pdf">GPT-4</a> scored 86% and 84% for MMLU and MATH respectively. The chart below shows examples of saturation across benchmarks</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!K2M5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb36ac303-da1e-4ce3-bab7-1e44b8c1b546_3400x2400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!K2M5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb36ac303-da1e-4ce3-bab7-1e44b8c1b546_3400x2400.png 424w, https://substackcdn.com/image/fetch/$s_!K2M5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb36ac303-da1e-4ce3-bab7-1e44b8c1b546_3400x2400.png 848w, https://substackcdn.com/image/fetch/$s_!K2M5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb36ac303-da1e-4ce3-bab7-1e44b8c1b546_3400x2400.png 1272w, https://substackcdn.com/image/fetch/$s_!K2M5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb36ac303-da1e-4ce3-bab7-1e44b8c1b546_3400x2400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!K2M5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb36ac303-da1e-4ce3-bab7-1e44b8c1b546_3400x2400.png" width="1456" height="1028" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b36ac303-da1e-4ce3-bab7-1e44b8c1b546_3400x2400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1028,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:578198,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!K2M5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb36ac303-da1e-4ce3-bab7-1e44b8c1b546_3400x2400.png 424w, https://substackcdn.com/image/fetch/$s_!K2M5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb36ac303-da1e-4ce3-bab7-1e44b8c1b546_3400x2400.png 848w, https://substackcdn.com/image/fetch/$s_!K2M5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb36ac303-da1e-4ce3-bab7-1e44b8c1b546_3400x2400.png 1272w, https://substackcdn.com/image/fetch/$s_!K2M5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb36ac303-da1e-4ce3-bab7-1e44b8c1b546_3400x2400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-11" href="#footnote-11" target="_self">11</a></p><p>Some of the best benchmarks that we have for scientific capabilities&#8212;<a href="https://arxiv.org/abs/2311.12022">GPQA diamond</a> and <a href="https://www.futurehouse.org/research-announcements/lab-bench-measuring-capabilities-of-language-models-for-biology-research">LAB-bench</a>&#8212;require the models to answer multiple-choice questions about the subject matter. These questions can be thoughtfully designed, but the conclusions from these tests are limited. They imply the systems were very good at answering a narrow set of scientific questions, it doesn&#8217;t imply very much about whether the models can <em>do the work. </em>For what it is worth, the AI systems clearly aren&#8217;t just multiple choice machines&#8212;systems <a href="https://x.com/hsu_steve/status/1835095080199504073">seem to be able to solve problems</a> from the famously difficult <em>Classical Electrodynamics </em>which can require days worth of effort from graduate-level physicists.</p><p>One approach to understanding the scientific capabilities of AI systems is to decompose the process of doing research into discrete steps, and evaluate each step piecemeal. For example, this paper tested <a href="https://arxiv.org/abs/2409.04109">hypothesis generation</a>, and found, "LLM-generated ideas are judged as more novel (p &lt; 0.05) than human expert ideas while being judged slightly weaker on feasibility&#8221;. The AI developers and the UK and US Safety Institutes are likely to maintain more comprehensive private taxonomies to track progress across the research pipeline, though there are strong incentives for these organisations not to release these.</p><p>Aside from benchmarks, it can be difficult to interpret progress clearly, because the people on the <em>very </em>frontier, who can see best where progress is headed, have been very heavily selected for conviction in the current technical paradigm. Sometimes, these researchers are modelled as capitalists, &#8216;hyping&#8217; the potential future capabilities to fundraise or make a project. This is an incomplete picture&#8212;no doubt there are people within AI labs who are &#8216;selling&#8217; the future, but there are also many researchers who very sincerely think that AI systems will have transformative capabilities. For example, the Chief AI Scientist at Anthropic, Jared Kaplan, who is also a Professor of Theoretical Physics at Johns Hopkins, <a href="https://youtu.be/4a5lzYreMME?si=Q4uarNroTo6GaH2z&amp;t=198">said in a talk at CERN</a>: &#8220;I think in certain respects AI systems are approaching human level performance. I think there are some challenges for continuing this progress but I think there aren&#8217;t any really compelling blockers for AI doing things like helping to automate science.&#8221; Likewise, John Schulman, the cofounder and former head of &#8216;post-training&#8217; at OpenAI, <a href="https://www.dwarkeshpatel.com/p/john-schulman">anticipates</a> that AI systems will match top AI researchers within 5 years, and Demis Hassabis, the CEO of Google DeepMind, <a href="https://www.thetimes.com/business-money/technology/article/ai-has-the-potential-to-cure-all-diseases-says-deepmind-chief-9hbdp5kpm">believes</a>, &#8220;we are in shooting distance of curing all diseases&#8221;. These claims should be treated with a healthy mixture of scepticism and seriousness, and they should not be dismissed out of hand.</p><p>This picture is made evermore difficult by the distortion of academia. Because AI developers pay salaries 10 times or more what can be earned in academia, there has been an enormous movement out of universities into OpenAI, Google DeepMind, and Anthropic. Those who are left behind are strongly selected from scepticism of the current technical paradigm.</p><p>The essential takeaway is that there have been very steep improvements in AI capabilities thus far, and as we will discuss in the next section, there are compelling reasons to believe this will continue. But exactly how this evolves, and where it might end is very uncertain. However, the most optimistic case&#8212;which is sincerely held by a number of people building on the frontier&#8212;is that this could usher in a period of explosive economic growth (~20% GDP growth or more).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-12" href="#footnote-12" target="_self">12</a> Though it is unlikely, it is within the realm of possibility.</p><h2>2. How AI became computationally intensive</h2><p>Some breakthrough technologies unlock our ability to exchange a latent resource for useful outputs. Steam engines are like this&#8212;pour the coal in, and the steam engine reliably converts the energy into the rotation of a crankshaft. This has all kinds of downstream uses: moving a train carriage, driving a pulley in a factory, or pumping water. Likewise, in the Haber-Bosch process, pour in hydrogen and nitrogen, and receive ammonia, which can be used as fertiliser. What makes these processes so powerful is their scalability: to move more goods before the steam engine, it would require adding more packhorses to carry goods along a turnpike road; before the Haber-Bosch process, growing more crops meant sending more boats to the Chincha Islands to harvest guano. With technology, there is dependable leverage&#8212;it is possible to continuously add more latent resources and receive back useful outputs.</p><p>Computational power is the unit of information processing, in brains and in computers. It would be incredibly useful if we could develop a technology which allowed us to pour in computational power and exchange this for useful informational processing capabilities. &#8216;Ordinary&#8217; computing does a version of this; but it is fragile and limited. Ordinary computers can only do processing tasks which have been specified by a program in advance, whereas AI systems have the capacity to learn. This is one intuition for what AI research is doing: it is finding reliable and scalable mechanisms&#8212;just like the steam engine or the Haber-Bosch reactor&#8212;that allows us to exchange processing power for flexible and adaptive intelligence.&nbsp;</p><p>This is unintuitive. <em>Prima facie, </em>AI research should be about having deep insights into the nature of intelligence, and designing machines that reflect these insights. Computer scientist Rich Sutton has called this realisation, <a href="http://www.incompleteideas.net/IncIdeas/BitterLesson.html">&#8216;The Bitter Lesson&#8217;</a>&#8212;what has driven AI progress is not the profound theories of researchers, but general strategies which allow AI systems to leverage greater amounts of computational power. In earlier approaches to AI, researchers would <a href="https://en.wikipedia.org/wiki/Hand_coding">&#8216;hand-code&#8217;</a> how they thought an AI system ought to learn, for example, what features of an image to recognise in order to classify the picture; but neural networks, the foundation of modern AI, allows the system to <em>learn for itself </em>what features are salient, through a training process.</p><h4>Increasing the amount of computational power during training</h4><p>The neural network is like a little computer which can be programmed by adjusting a series of dials. The aim of a neural network is to predict an output given a set of inputs. The iterative process of tuning these dials to improve the prediction is called &#8216;training&#8217;. The people creating the network supervise the training process by showing the data and the answers, but crucially, it doesn&#8217;t involve telling the network <em>how </em>it ought to process and understand the image. In other words, our process of trial and improvement tweaking of dials is essentially letting the little computer, by itself, search for the best way it can be programmed to achieve its goal, unlike ordinary computers which need a human to figure out a program first and then somehow communicate it to the computer. Dario Amodei <a href="https://www.dwarkeshpatel.com/p/dario-amodei">described</a> the training process in this way:</p><blockquote><p><em>&#8220;You [the AI researcher] get the obstacles out of their way. You give them good data, you give them enough space to operate in, you don't do something stupid like condition them badly numerically [i.e. tweak the dials poorly], and they want to learn. They'll do it.&#8221;</em></p></blockquote><p>What your neural network will end up learning depends on the goal you give them to pursue, and in what you ask the network to predict. There&#8217;s two ways it is possible to do this:</p><ol><li><p>Directly optimise for a specific capability.</p></li><li><p>Optimise for a <em>related </em>goal, and hope that important capabilities emerge downstream.</p></li></ol><p>Large language models, the basis of recent progress in AI, take the second approach. A language model is optimised to predict the next word<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-13" href="#footnote-13" target="_self">13</a> in a sequence, based on the words that have come before. Google DeepMind&#8217;s Gemini 1.5 model is able to take in <a href="https://arxiv.org/pdf/2403.05530">nearly one million words of input</a>, to give the most subtle and accurate prediction of the next word of output. This isn&#8217;t an <em>inherently </em>valuable task in the way that a neural network which is trained to predict whether a dot on the screen is a tumour or a harmless cyst <em>is </em>inherently useful.</p><p>While predicting the next word isn&#8217;t inherently useful, <em>pursuing this goal</em> is still enormously powerful. Text has a very rich structure, meaning the words aren&#8217;t randomly assorted: whoever wrote them chose their order to convey an idea. If a neural network can understand that structure&#8212;say, by parsing all of human knowledge in the training process&#8212;the network has a representation of how everything fits together. Researchers would call this a &#8216;world model&#8217;.</p><p><em>Prediction </em>is a meaningful task&#8212;if the network can take most of a book as input, and predict the end without having seen the book before, there is some sense in which it understands the content.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-14" href="#footnote-14" target="_self">14</a></p><p>The measure of how far these predictions are from reality is known as the &#8216;training loss&#8217;. There is an <em>extremely</em> predictable relationship between the model&#8217;s size, the amount of training data it uses, the amount of computational power (compute) it is trained on; and the model&#8217;s loss. As the models get bigger, and are trained on more data, using more compute; its loss declines. In other words, its predictions get better. Amodei <a href="https://www.dwarkeshpatel.com/p/dario-amodei">has noted</a> the declines are, &#8220;sometimes predictable even to several significant figures which you don't see outside of physics.&#8221;&nbsp;</p><p>What is much less predictable is whether these declines in training loss translate into useful capabilities. It has been particularly surprising how well improvements on next token prediction <em>have </em>converted<em> </em>into useful capabilities so far. In expectation of these capabilities, AI developers have scaled their models dramatically: Google DeepMind&#8217;s Gemini 1.5, released in December 2023, <a href="https://epoch.ai/blog/training-compute-of-frontier-ai-models-grows-by-4-5x-per-year#language-models-caught-up-to-the-frontier-around-2020">was trained using 6.7 million times the amount of compute</a> used to train a state-of-the-art large language model in June 2017. To put this in <a href="https://epoch.ai/blog/training-compute-of-frontier-ai-models-grows-by-4-5x-per-year#language-models-caught-up-to-the-frontier-around-2020">broader perspective</a>, since 2012 the amount of computation power used to train the largest models has grown by 100 million-fold.</p><h4>Using more compute while the model is running</h4><p>Thus far, we have only described the opportunity to pour in additional compute <em>during the training process</em> to receive useful outputs. It is also desirable to pour in additional compute while the model is being used; technically called &#8216;inference&#8217;. A paradigm example of an AI system to take advantage of &#8216;inference-time compute&#8217; was DeepMind&#8217;s <a href="https://www.nature.com/articles/nature16961">AlphaGo</a>. AlphaGo was trained to play the board game <a href="https://en.wikipedia.org/wiki/Go_(game)">Go</a>, or, <em>to be specific</em>, it was trained to find the next move from a given board state which maximised win probability. One way the system <em>might </em>have solved this problem is to imitate the moves of a human expert. Indeed, this was the initial approach to training. But a more advanced way of learning would be for the model to play against itself predicting what could happen in future moves and learn from that guess. This kind of planning could also be used when the model was competing. When DeepMind&#8217;s system was allowed to use this additional capability before choosing a move, performance jumped from 1500 <a href="https://en.wikipedia.org/wiki/Elo_rating_system">ELO</a> to 3000 ELO.</p><p>AlphaGo was able to exchange compute at inference-time for a jump in capabilities because the game of Go is well suited to running a form of <a href="https://en.wikipedia.org/wiki/Monte_Carlo_tree_search">tree search</a>. It has a very clear goal (to win the game) and it has a reasonably constrained search space (the potential moves on the Go board). Determining the best next step is relatively straightforward.&nbsp;</p><p>By contrast, &#8216;language space&#8217; isn&#8217;t like this. There are many prompts which do not have a formally &#8216;correct&#8217; response&#8212;determining what is &#8216;better&#8217; is much more subjective. Also, the &#8216;space&#8217; of potential directions in language is much larger than the potential next moves on a go board, especially as the responses get much longer. It would be very difficult for a tree search procedure for language to know what direction to search in. Because of this, when ChatGPT was released in November 2022, it had been trained on 500 billion tokens (think: words) but it could only use 4,096 tokens to respond to each prompt. It couldn&#8217;t use more compute while it was running to generate better answers; it was hobbled.&nbsp;</p><p>This changed in September 2024. OpenAI released &#8216;o1&#8217;, a new series of models which generate &#8216;Chains of Thought&#8217; before responding to a prompt. Chains of thought allow the model to use additional processing in response to more difficult questions. In the <a href="https://openai.com/index/learning-to-reason-with-llms/">o1 release blog post</a>, OpenAI showed a graph that demonstrates how <em>as the compute budget is expanded</em>, its performance on a qualification test for the US Maths Olympiad improves.&nbsp;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!q3Uz!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8157e8a1-b954-434c-9745-280b0df889c5_804x804.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!q3Uz!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8157e8a1-b954-434c-9745-280b0df889c5_804x804.png 424w, https://substackcdn.com/image/fetch/$s_!q3Uz!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8157e8a1-b954-434c-9745-280b0df889c5_804x804.png 848w, https://substackcdn.com/image/fetch/$s_!q3Uz!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8157e8a1-b954-434c-9745-280b0df889c5_804x804.png 1272w, https://substackcdn.com/image/fetch/$s_!q3Uz!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8157e8a1-b954-434c-9745-280b0df889c5_804x804.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!q3Uz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8157e8a1-b954-434c-9745-280b0df889c5_804x804.png" width="804" height="804" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8157e8a1-b954-434c-9745-280b0df889c5_804x804.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:804,&quot;width&quot;:804,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!q3Uz!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8157e8a1-b954-434c-9745-280b0df889c5_804x804.png 424w, https://substackcdn.com/image/fetch/$s_!q3Uz!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8157e8a1-b954-434c-9745-280b0df889c5_804x804.png 848w, https://substackcdn.com/image/fetch/$s_!q3Uz!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8157e8a1-b954-434c-9745-280b0df889c5_804x804.png 1272w, https://substackcdn.com/image/fetch/$s_!q3Uz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8157e8a1-b954-434c-9745-280b0df889c5_804x804.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The important takeaway from this section is that AI developers are constantly improving their&nbsp; understanding of how using more computational power can increase the useful capabilities of the models. Granted, the current methods for applying additional compute might stop having returns, but it is likely that the future methods we find will continue to follow this pattern.</p><p>In the current world, intelligence is scarce and special&#8212;just as fertiliser or mechanical power was before the steam engine. It is weird to say, but creating more intelligence today is laborious: it occurs only in humans who take decades to mature and require lots of education. In the not too distant future, intelligence equal to, or beyond human-level, will be constrained <em>only </em>by our ability to pour in computer chips and electricity.</p><h2>3. Further progress and deployment of AI systems will use 10&#8217;s to 100&#8217;s of gigawatts</h2><p>Let us begin with historical trends. Since 2012, the amount of computational power used to train the largest models has increased 100 million-fold: how have we done this? There are <a href="https://epochai.org/blog/trends-in-machine-learning-hardware#:~:text=FLOP%2Fs%20performance%20in%2047,bandwidth%20doubled%20every%204%20years.">beneficial tailwinds</a>: the computational power of the state-of-the-art AI chip has doubled every 2.3 years, as shown in the chart below; and the energy efficiency of hardware has doubled every 3 years.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-15" href="#footnote-15" target="_self">15</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!O-Ti!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb4424008-9a5e-402a-a57f-168c360d3a00_1600x1000.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!O-Ti!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb4424008-9a5e-402a-a57f-168c360d3a00_1600x1000.png 424w, https://substackcdn.com/image/fetch/$s_!O-Ti!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb4424008-9a5e-402a-a57f-168c360d3a00_1600x1000.png 848w, https://substackcdn.com/image/fetch/$s_!O-Ti!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb4424008-9a5e-402a-a57f-168c360d3a00_1600x1000.png 1272w, https://substackcdn.com/image/fetch/$s_!O-Ti!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb4424008-9a5e-402a-a57f-168c360d3a00_1600x1000.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!O-Ti!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb4424008-9a5e-402a-a57f-168c360d3a00_1600x1000.png" width="1456" height="910" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b4424008-9a5e-402a-a57f-168c360d3a00_1600x1000.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:910,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!O-Ti!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb4424008-9a5e-402a-a57f-168c360d3a00_1600x1000.png 424w, https://substackcdn.com/image/fetch/$s_!O-Ti!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb4424008-9a5e-402a-a57f-168c360d3a00_1600x1000.png 848w, https://substackcdn.com/image/fetch/$s_!O-Ti!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb4424008-9a5e-402a-a57f-168c360d3a00_1600x1000.png 1272w, https://substackcdn.com/image/fetch/$s_!O-Ti!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb4424008-9a5e-402a-a57f-168c360d3a00_1600x1000.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-16" href="#footnote-16" target="_self">16</a></p><p>However, the increase in training compute exceeds the rate of hardware improvements: since 2010, training compute has doubled every six months! Language models have moved faster still: doubling happens roughly every five months, shown in the graph below.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!stqe!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fadbbb908-1fba-4c30-b720-000f7b1cd772_1600x1000.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!stqe!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fadbbb908-1fba-4c30-b720-000f7b1cd772_1600x1000.png 424w, https://substackcdn.com/image/fetch/$s_!stqe!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fadbbb908-1fba-4c30-b720-000f7b1cd772_1600x1000.png 848w, https://substackcdn.com/image/fetch/$s_!stqe!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fadbbb908-1fba-4c30-b720-000f7b1cd772_1600x1000.png 1272w, https://substackcdn.com/image/fetch/$s_!stqe!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fadbbb908-1fba-4c30-b720-000f7b1cd772_1600x1000.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!stqe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fadbbb908-1fba-4c30-b720-000f7b1cd772_1600x1000.png" width="1456" height="910" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/adbbb908-1fba-4c30-b720-000f7b1cd772_1600x1000.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:910,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!stqe!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fadbbb908-1fba-4c30-b720-000f7b1cd772_1600x1000.png 424w, https://substackcdn.com/image/fetch/$s_!stqe!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fadbbb908-1fba-4c30-b720-000f7b1cd772_1600x1000.png 848w, https://substackcdn.com/image/fetch/$s_!stqe!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fadbbb908-1fba-4c30-b720-000f7b1cd772_1600x1000.png 1272w, https://substackcdn.com/image/fetch/$s_!stqe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fadbbb908-1fba-4c30-b720-000f7b1cd772_1600x1000.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-17" href="#footnote-17" target="_self">17</a></p><p>While using a computer at home is not at all energy-intensive, the kind of computation that AI systems are doing is very energy-demanding. A state-of-the-art chip, the H100 Graphical Processing Unit (GPU) made by NVIDIA, has an annual power draw 44% higher than the average UK resident!<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-18" href="#footnote-18" target="_self">18</a> This is set to increase, in the next generation chip (the B200) to 150% more power than the average UK resident. The exact amount of power that Google DeepMind, Anthropic, and OpenAI use to train their systems is kept secret, for competitive reasons, but Meta published a report with their latest model release in July 2024, which noted that training their largest model <a href="https://scontent-lhr8-1.xx.fbcdn.net/v/t39.2365-6/463020162_522238820565582_8192401983671993921_n.pdf?_nc_cat=108&amp;ccb=1-7&amp;_nc_sid=3c67a6&amp;_nc_ohc=6V_W4zoVlq0Q7kNvgHV6_Gk&amp;_nc_zt=14&amp;_nc_ht=scontent-lhr8-1.xx&amp;_nc_gid=ATFiT-sFUQpdhWuVQNOzvPZ&amp;oh=00_AYBeYjCOR6Cacf8E7_9-VHm43tEwXPvebkElCLjnnvHmUw&amp;oe=673BE619">used 16,000 active GPUs</a>. EpochAI, a third-party research organisation, <a href="https://epochai.org/blog/can-ai-scaling-continue-through-2030">estimates</a> this required 27 megawatts (MW) of installed electricity capacity.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-19" href="#footnote-19" target="_self">19</a> This is approximately the power supply required for 88,000 UK households, which more than the number of households in York.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-20" href="#footnote-20" target="_self">20</a></p><p>One useful intuition is that each AI chip has the power demand of a person, or maybe soon a household, and each AI datacentre has the power demand of a small city. The large datacentre campuses will be similar to the largest cities. Of course, it is quite remarkable how much power it takes to train the current AI systems, but the steep trendlines point towards this amount of computational power and energy becoming quite small, quite quickly!&nbsp;</p><h4>What is the future of AI datacentre and energy demand?</h4><p>A <a href="https://situational-awareness.ai/racing-to-the-trillion-dollar-cluster/">report by a former OpenAI employee</a>, Leopold Aschenbrenner, extrapolated the current trendlines in computational power increases, and noted that current growth rates imply:</p><ul><li><p>The largest model in 2026 will be trained on the equivalent computational power of one million of today&#8217;s state-of-the-art GPUs and require 1 gigawatt (GW) of power. (Of course, hardware improvements mean it will be a smaller number of more intensive chips, so hereafter we&#8217;ll use the unit H100-equivalent for comparison.)</p></li><li><p>In 2028, the largest model will use 10 million H100-equivalents of computational power, and 10GW of electricity.</p></li><li><p>In 2030, this will jump to 100 million H100-equivalents and 100 GW of electricity.</p></li></ul><p>This is daunting. The 16,000 H100s which Meta used to train their most recent model, which required the same power as York, looks microscopic in comparison to the figures for the simple extrapolations for the end of the decade. Were this trendline to hold, and the length of training runs <a href="https://epochai.org/blog/the-longest-training-run">reaches optimum</a>, <strong>the single largest model in 2030 would require 2.15 times more power than the UK&#8217;s entire electricity generation in 2021.</strong><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-21" href="#footnote-21" target="_self">21</a><strong> </strong>Of course, this is not a prediction, merely an observation of what continuing straight lines would imply.&nbsp;</p><p>Thus far, we have only described trends in training compute, not in the inference of systems (i.e. when the models are being run). If the inference-time compute paradigm which OpenAI have developed using chains-of-thought can be extended further, the computational intensity of inference will rise sharply. This will be compounded by increased <em>frequency </em>of model inference. As AI systems become more integrated into the economy, inference will become the dominant form of AI computing by far. Jensen Huang, the CEO of NVIDIA, <a href="https://x.com/RihardJarc/status/1845453408557289653">expects</a> the amount of inference to go up &#8216;by a billion times&#8217; (and given the 100 million times increase in training compute in the last decade, we take this estimate seriously!)</p><p>SemiAnalysis <a href="https://semianalysis.com/2024/03/13/ai-datacenter-energy-dilemma-race/">published</a> an estimate in March 2024 that US &#8216;AI Data Centre Critical IT Power&#8217; will rise to 56.3 GW in 2028, from 8.5 GW in 2024. Globally, they <a href="https://semianalysis.com/2024/03/13/ai-datacenter-energy-dilemma-race/">expect</a> AI datacentre power demand to rise by approximately 40 GW by 2026. These trends provide an important indicator about the criticality of this current moment: it is very much the beginning of the buildout.</p><p>The growth in AI datacentres is constrained by energy availability. There is not 40GW of spare energy capacity around the world, and so the cloud providers are taking steps to meet their power demands. Cloud providers have snapped up the limited amount of spare capacity it was possible to buy&#8212;for example, Amazon has bought a <a href="https://www.ans.org/news/article-5842/amazon-buys-nuclearpowered-data-center-from-talen/">960 MW nuclear reactor</a>, and Microsoft has signed a 20-year power purchase agreement with Constellation Energy to reopen <a href="https://www.bloomberg.com/news/articles/2024-09-25/microsoft-to-pay-hefty-price-for-three-mile-island-clean-power">an 836 MW reactor at Three Mile Island</a><strong>. </strong>SemiAnalysis <a href="https://www.semianalysis.com/p/100000-h100-clusters-power-network">reports</a>, <em>&#8220;[T]he search for power is so dire, X.AI is even converting an old factory in Memphis Tennessee into a datacenter due to the lack of other options.&#8221;</em> As an indicator of the intensity of the buildout: to create the power supply for this datacentre, x.ai&#8230;</p><blockquote><p><em>&#8220;[P]ut a bunch of mobile [natural gas] generators usually reserved for natural disasters outside, add[ed] a Tesla battery pack, [drove] as much power as we can from the grid, tap[ped] the natural gas line that's going to the natural gas plant two miles away, the gigawatt natural gas plant&#8230;[and got] a cluster built as fast as possible."</em></p></blockquote><p>This project was <a href="https://www.businessinsider.com/jensen-huang-elon-musk-supercomputer-xai-grok-2024-10">completed in 19 days</a>, despite the fact that constructing a 100,000 GPU cluster ordinarily takes a year. (It also ordinarily costs <a href="https://www.dwarkeshpatel.com/p/dylan-jon?open=false#%C2%A7being-head-of-compute-at-an-ai-lab">$1 billion, but they were willing to spend $4 to $5 billion</a>.)</p><p>To fuel further growth, the cloud providers are enabling the construction of new energy assets. Oracle has a <a href="https://www.newcivilengineer.com/latest/software-company-oracle-plans-to-use-three-smrs-to-power-ai-focused-data-centre-23-09-2024/">permit to build three SMRs</a> and Google announced a partnership which will give them <a href="https://blog.google/outreach-initiatives/sustainability/google-kairos-power-nuclear-energy-agreement/">seven SMRs to provide 500 MW for datacentres</a>, starting in 2030. Amazon has <a href="https://www.aboutamazon.com/news/sustainability/amazon-nuclear-small-modular-reactor-net-carbon-zero">partnered with Dominion Energy to build SMRs for datacentres</a>. Most ambitiously, <a href="https://www.bloomberg.com/news/articles/2024-09-24/openai-pitched-white-house-on-unprecedented-data-center-buildout">OpenAI asked the Biden Administration</a> to construct between five and seven 5GW datacentre campuses across the US.<br><br>The exact level of capital expenditure on AI infrastructure (let alone on energy generation assets) is difficult to disaggregate from the earnings reports of big tech companies. <a href="https://www.economist.com/business/2024/07/28/what-could-kill-the-1trn-artificial-intelligence-boom">This estimate suggests</a> big tech companies will spend more than $100 billion on AI infrastructure in 2024, and SemiAnalysis estimates that Microsoft will independently spend <a href="https://www.semianalysis.com/i/144399864/is-microsoft-even-committed">more than $50 billion.</a> Estimates of future capital expenditure vary from hundreds of billions to <a href="https://x.com/MasaSonCap/status/1851267939820798015">nine trillion dollars</a>.</p><p>As with many things in AI, it is uncertain, but likely to be big.</p><h2>4. The UK would power AI datacentres using nuclear power, not wind and solar</h2><p>We investigated which source of power generation would be most suitable for AI datacentres in the UK, to determine where reforms should be focused. We compared two forms of low-carbon energy&#8212;nuclear power, or a blend of wind, solar, batteries backup, and natural gas reserve. We didn&#8217;t consider the possibility of using entirely natural gas to power AI datacentres, as we considered this incompatible with emissions aims, though it is a potential approach.</p><p>It is important to note that not only<em> </em>do AI datacentres need lots of power, but they need incredibly reliable power. Datacentres require high uptime&#8212;their service level agreements typically stipulate &#8220;five nines of reliability&#8221;, meaning the datacentre can have downtime of 5 minutes and 15 seconds over the course of a year. This effectively dictates the power cannot fail, as, <a href="https://www.future-tech.co.uk/data-centre-availability-and-reliability-an-explanation-suggested-kpis/#:~:text=It%20is%20generally%20accepted%20though,several%20hours%20or%20even%20days.">this report</a> notes &#8220;even a 25 millisecond power outage could take down the entire datacentre for several hours or even days&#8221;.</p><p>AI datacentres need reliability for two main reasons:</p><ol><li><p>The current technical regime for training large models currently requires synchronisation between compute assets. If an AI datacentre is contributing to a large training run, and it goes offline during a training step, it will disrupt the whole training process.</p></li><li><p>Second, despite the power intensity of GPUs, electricity only makes up <a href="https://semianalysis.com/2023/12/04/gpu-cloud-economics-explained-the/">a small fraction of the &#8216;Total Cost of Ownership&#8217;</a>: the monthly GPU server hosting costs are just $1872, while server capital costs are $7026. Therefore, having GPUs sit idle is much more expensive than building redundancy.</p></li></ol><h4>The renewables, batteries, and gas blend is impractically expensive, and undesirably polluting and land intensive.</h4><p>To think about how the optimal blend works; it makes sense to layer on the tradeoffs. First, wind and solar are highly intermittent, and so getting consistent output requires building much more than would be naively estimated. (A 1GW solar plant will only actually give you 1GW in the most fleetingly intense moments of the summer, most of the time it will fall below.) Next, because it is sometimes windy when it is not sunny, and sunny when it is not windy, we say that wind and solar have covariance, and so there is always value in having some of each in the mix. Solar and wind are roughly as cheap, so cost is not a major determinant of which to choose.&nbsp;</p><p>There will be moments throughout the day when it is neither sunny nor windy. For these it is necessary to employ batteries. This does not fully solve the problem, because sometimes there will be low wind speeds and clouds for a week, and while this could be solved by building complete battery backup, it becomes enormously expensive. Batteries are cost-competitive if they are being used constantly (roughly every four hours), but they are expensive to sit idle, and so while batteries are efficient for overnight use, it is impractical to use them for a longer period. For extended periods where wind speeds are low and it is cloudy, it is necessary to use a natural gas backup.</p><p>We estimated that the cost-minimising way to use a blend of solar, wind, batteries, and natural gas; for every 1 GW of stable output, it would require 8 GW of solar panels, 0.37 GW of wind power, 12 GWh of battery-backup, complete gas-backup, and LNG import capacity of 451 million cubic metres of gas. (Our method is included in the <a href="https://inferencemagazine.substack.com/getting-ai-datacentres-in-the-uk#Technical-Appendix">Technical Appendix</a>.)</p><p>This presents a number of challenges:</p><ol><li><p><strong>The land use.</strong> Building 8GW of installed solar and 0.37GW of wind capacity would require 160km<sup>2</sup> and 41km<sup>2</sup> respectively. (We assume the wind is onshore, as costs offshore are harder to predict as getting energy back to shore is more difficult).This means that <em>each gigawatt </em>would require more than 200km<sup>2</sup> of contiguous land. As a point of reference, Cardiff is 140km<sup>2</sup> and Reading is 40km<sup>2</sup>. This scales very poorly. Some datacentre campuses are much larger than one gigawatt, for example, imagine that <a href="https://www.theinformation.com/articles/microsoft-and-openai-plot-100-billion-stargate-ai-supercomputer">Microsoft and OpenAI&#8217;s 5GW datacentre campus</a> was powered by renewables in the UK, it would require over 1000km<sup>2</sup> of contiguous land.</p></li><li><p><strong>The land would need to be next to LNG import capacity. </strong>The UK has three LNG import terminals, one in Kent and two in South Wales. The natural gas generation terminal would need to be some distance from population centres to reduce air pollution, but also close enough to the import terminal that gas pipelines are not necessary. It is either necessary to find ways to build these generation facilities in Kent or South Wales <em>while being contiguous </em>with the solar and wind farms, or it might be necessary to build a new LNG import terminal, but neither approach scales well, and the latter would add substantially to costs.&nbsp;&nbsp;</p></li><li><p><strong>Emissions.</strong> As natural gas would generate 28% of the electricity, over its expected life cycle, the renewable blend would produce 40% more carbon emissions than equivalent nuclear capacity.</p></li><li><p><strong>Cost.</strong> The levelised cost would be &#163;106/MWh. This would be internationally uncompetitive for creating an AI datacentre&#8212;the equivalent cost for Texas to produce energy with the same carbon footprint would be &#163;74/MWh. Should emissions be desired to fall further, the output electricity would only become even more uncompetitively expensive.&nbsp;</p></li><li><p><strong>Safety. </strong>Even when natural gas is only used 28% of the time, the mix leads to <a href="https://ourworldindata.org/safest-sources-of-energy">27 times more &#8216;expected deaths&#8217; per TWh</a> (mostly from air pollution), when compared to the risk of accidents with nuclear power.</p></li></ol><p>As a reminder, these numbers are for 1GW of consistent output, and global datacentre buildout is expected to be around 40 GW by 2026.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-22" href="#footnote-22" target="_self">22</a> It is probable to expect 10s of gigawatts of power demand in the coming years. Renewable approaches will not be able to scale to this level, and to provide the concentration of power required onto the largest datacentre campuses.</p><h4>Nuclear is more suitable because it is reliable and safe</h4><p>Contrary to wind and solar, nuclear power has very stable output, which is ideal for meeting the &#8216;five nines&#8217; requirement that datacentres have &#8211; 99.999% reliability.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-23" href="#footnote-23" target="_self">23</a> The US has a nuclear <a href="https://www.energy.gov/ne/articles/nuclear-power-most-reliable-energy-source-and-its-not-even-close#:~:text=Nuclear%20Has%20The%20Highest%20Capacity%20Factor&amp;text=This%20basically%20means%20nuclear%20power,than%20wind%20and%20solar%20plants.">capacity factor of 92.5%</a> as a benchmark. The chart below shows US total energy generation in a week in March last year; note that nuclear is the green block at the bottom&#8212;extremely stable!&nbsp;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!iiES!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d4c1702-82d1-4b43-a1ae-2dfeb9d6cf4f_1600x838.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!iiES!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d4c1702-82d1-4b43-a1ae-2dfeb9d6cf4f_1600x838.png 424w, https://substackcdn.com/image/fetch/$s_!iiES!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d4c1702-82d1-4b43-a1ae-2dfeb9d6cf4f_1600x838.png 848w, https://substackcdn.com/image/fetch/$s_!iiES!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d4c1702-82d1-4b43-a1ae-2dfeb9d6cf4f_1600x838.png 1272w, https://substackcdn.com/image/fetch/$s_!iiES!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d4c1702-82d1-4b43-a1ae-2dfeb9d6cf4f_1600x838.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!iiES!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d4c1702-82d1-4b43-a1ae-2dfeb9d6cf4f_1600x838.png" width="1456" height="763" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3d4c1702-82d1-4b43-a1ae-2dfeb9d6cf4f_1600x838.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:763,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!iiES!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d4c1702-82d1-4b43-a1ae-2dfeb9d6cf4f_1600x838.png 424w, https://substackcdn.com/image/fetch/$s_!iiES!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d4c1702-82d1-4b43-a1ae-2dfeb9d6cf4f_1600x838.png 848w, https://substackcdn.com/image/fetch/$s_!iiES!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d4c1702-82d1-4b43-a1ae-2dfeb9d6cf4f_1600x838.png 1272w, https://substackcdn.com/image/fetch/$s_!iiES!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d4c1702-82d1-4b43-a1ae-2dfeb9d6cf4f_1600x838.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-24" href="#footnote-24" target="_self">24</a></p><p>Furthermore, there has been an international realisation that nuclear power is safe, clean, and necessary to meet climate goals. 20 countries at COP 28&#8212;including the US and UK&#8212;<a href="https://www.energy.gov/articles/cop28-countries-launch-declaration-triple-nuclear-energy-capacity-2050-recognizing-key">announced their intention to triple global nuclear</a> energy capacity by 2050. The world&#8217;s biggest investment banks have <a href="https://www.ft.com/content/96aa8d1a-bbf1-4b35-8680-d1fef36ef067">announced their intention to finance this aim</a>, <a href="https://x.com/TheMongrel_Cat/status/1844726286024937925">Italy</a> and <a href="https://x.com/TheMongrel_Cat/status/1842918133763617108">India</a> announced plans to accelerate the construction of nuclear power, and <a href="https://world-nuclear.org/information-library/country-profiles/countries-g-n/japan-nuclear-power">Japan is looking to reopen 13 nuclear reactors</a>. Since then, the US has said it will <a href="https://www.datacenterdynamics.com/en/news/biden-admin-details-roadmap-to-triple-us-nuclear-power-by-2050-add-200gw/">add 200 GW of nuclear power by 2050</a>.</p><p>A negative perception of nuclear energy has come from its association to nuclear weapons, and the infrequent-but-visible reactor meltdowns and subsequent evacuations. The Pripyat Ferris Wheel and empty swimming pool have a prominent effect on our collective psyche towards nuclear energy safety, but we have no similar association for the Banqiao Dam Collapse which killed 171,000 people in 1975.&nbsp;</p><p>When compared to other sources of energy, especially practical stable generators alternatives, nuclear power is comparatively much safer and cleaner per TWh of generation.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!WDh0!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feda2e02f-f133-40d2-90c1-0bf1a848b483_1600x861.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!WDh0!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feda2e02f-f133-40d2-90c1-0bf1a848b483_1600x861.png 424w, https://substackcdn.com/image/fetch/$s_!WDh0!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feda2e02f-f133-40d2-90c1-0bf1a848b483_1600x861.png 848w, https://substackcdn.com/image/fetch/$s_!WDh0!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feda2e02f-f133-40d2-90c1-0bf1a848b483_1600x861.png 1272w, https://substackcdn.com/image/fetch/$s_!WDh0!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feda2e02f-f133-40d2-90c1-0bf1a848b483_1600x861.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!WDh0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feda2e02f-f133-40d2-90c1-0bf1a848b483_1600x861.png" width="1456" height="784" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/eda2e02f-f133-40d2-90c1-0bf1a848b483_1600x861.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:784,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!WDh0!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feda2e02f-f133-40d2-90c1-0bf1a848b483_1600x861.png 424w, https://substackcdn.com/image/fetch/$s_!WDh0!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feda2e02f-f133-40d2-90c1-0bf1a848b483_1600x861.png 848w, https://substackcdn.com/image/fetch/$s_!WDh0!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feda2e02f-f133-40d2-90c1-0bf1a848b483_1600x861.png 1272w, https://substackcdn.com/image/fetch/$s_!WDh0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feda2e02f-f133-40d2-90c1-0bf1a848b483_1600x861.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-25" href="#footnote-25" target="_self">25</a></p><p>There have been three high-profile reactor meltdowns: Three Mile Island (1979), Chernobyl (1986) and Fukushima (2011). The meltdown at Three Mile Island caused no deaths either directly or indirectly, and the radiation exposure for 2.2 million people who lived near to the New Jersey plant was, &#8220;approximately the same radiation dose as flying from New York to Los Angeles&#8221;.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-26" href="#footnote-26" target="_self">26</a></p><p>Our World in Data, an independent research organisation, has <a href="https://ourworldindata.org/what-was-the-death-toll-from-chernobyl-and-fukushima">reviewed the death tolls for Chernobyl and Fukushima</a>. Their literature review estimated the Chernobyl meltdown caused between 300 and 500 deaths; 30 direct deaths, and the remainder were indirect. At Fukushima in 2011, there were no direct deaths in the disaster. There were 40 to 50 injuries, and 7 years after the accident, it was reported that one worker died from lung cancer caused by radiation exposure at the event. However, there was a mass evacuation, which is estimated to have caused 2,313 deaths, from the physical or mental exertion of evacuation (from care homes and similar places). Disentangling which of these deaths were attributable to the evacuation following the meltdown, compared with the wider impact of the earthquake and tsunami, is necessarily difficult.</p><p>There is a particular dissonance between attitudes to fossil fuels and nuclear power. Unlike nuclear power, fossil fuels are continuously and gradually reducing the life expectancy of billions of people, but there is never a discrete moment where this is felt more acutely. <a href="https://thoughtscapism.com/2019/10/10/what-level-of-risk-justifies-denying-people-their-homes-a-look-at-fukushima-vs-pollution-in-big-cities/">This report</a> notes that, &#8220;moving to Tokyo would triple the populations&#8217; increase in risk of death [because of air pollution], compared to moving them back to the remaining off-limits zones in Fukushima.&#8221; <a href="https://conference.nber.org/conf_papers/f205791.pdf">This paper</a> estimates that the slowdown in nuclear power construction following the Chernobyl meltdown caused the loss of 33 million expected life years in the UK alone, or roughly 400,000 people, because of particulate poisoning.</p><p>To summarise, the UK will not be able to produce cost-competitive renewable energy for AI datacentres, that scales to the 10s of gigawatts required. However, nuclear power has all the necessary attributes&#8212;it is cleaner, safer, has more reliable generation&#8212;and as we <a href="https://inferencemagazine.substack.com/i/151677344/the-uk-should-create-special-compute-zones">discuss in a further section</a>, could become internationally cost competitive too.</p><h2>5. Going without AI datacentres would be a mistake</h2><p>A sceptical line of argument might say that while AI progress is happening; and AI datacentres and power will grow dramatically; and nuclear power will be the dominant energy source; it does not necessarily follow that the UK should be concerned with hosting AI datacentres. The strongest form of this argument claims: the pattern of economic history is that the gains from general-purpose technologies tend to come from adoption, and perhaps UK residents and businesses could buy access to AI datacentres internationally, while the UK could focus on the &#8216;highest value&#8217; parts of the AI value chain.</p><p>This argument is not enough&#8212;going without AI datacentres would be a mistake. The economic doctrine that the UK can sit atop the value chain, and selectively choose to engage with &#8216;high value&#8217; industries has led to a hollowing out of industry, and left the UK without growth. Hosting AI datacentres will enhance the UK&#8217;s economic security, allows directed growth into former industrial areas, and enables future frontier growth.</p><h4>A large fraction of the UK&#8217;s capital stock will be in AI datacentres</h4><p>It is very likely that computational power becomes a critical input into the production of goods and services, in a manner similar to energy. Just as the venture capitalist Marc Andreessen commented in 2011, &#8216;software is eating the world&#8217;, the information processing capabilities of AI systems will become tightly knit into all existing business processes and future ones. This will be especially true for the UK&#8217;s professional services exports. As a result, a very large fraction of the UK&#8217;s capital stock will be created, stored, and run in, or at least depend upon, AI datacentres. The UK will want these AI datacentres to be here, rather than overseas, connected through an undersea cable, to ensure it can protect these assets.</p><p>Furthermore, <em>precisely because</em> the gains from AI come from adoption, the UK needs to ensure access to computational power. As demand for computational power rises globally, it could be the case that UK businesses are unable to access the computational power they need. Right now, the Microsoft CFO said on <a href="https://www.microsoft.com/en-us/Investor/events/FY-2025/earnings-fy-2025-q1">a recent earnings call</a> that revenue growth in their cloud business was 33% but, &#8220;[d]emand continues to be higher than our available capacity.&#8221;</p><p>For decades, the UK decided to go without energy self-sufficiency. It would be imprudent to repeat the same mistake for computational power.</p><h4>Directing growth to former industrial areas</h4><p>The work to adopt and develop AI applications is likely to centre around London. The <a href="https://www.gov.uk/government/publications/artificial-intelligence-sector-study-2023/artificial-intelligence-sector-study-2023">Government&#8217;s AI Sector Study</a> shows that 75% of UK AI companies are based in London, the South East, or East of England. AI application developers are likely to agglomerate here because London has the best venture capital ecosystem and AI talent density outside San Francisco. Furthermore, adoption of AI systems is likely to focus on automation of business processes in professional service domains and scientific endeavours initially, which is also likely to begin within the &#8216;Golden Triangle&#8217;.</p><p>AI datacentres do not depend on network agglomeration in the same way. The construction and operation of these datacenters can be much more readily directed, to areas outside the Golden Triangle with strong industrial traditions. Very rarely does the opportunity of investment which is location-independent come along. There can be thousands of skilled jobs in the nuclear and datacentre construction and operation industries.&nbsp;&nbsp;</p><p>Furthermore, there are spillover benefits from developing these industries. UK workers will learn how to build nuclear reactors from the Korean Electric Power Corporation, who construct reactors for 25% of the price; or how to build AI datacentres with a Power Usage Effectiveness of 1.1, from Google.</p><h4>If the UK wants to participate in the frontier growth of the future</h4><p>The UK has enormous strengths in science, through its world-leading universities and research institutes. The process of research and development is likely to be transformed by AI systems in the coming years, and so maintaining the UK&#8217;s scientific advantage and the prospect of future growth it offers is likely to require differential integration of AI systems into the research cycle. It seems very likely that future UK growth will be downstream AI-enabled scientific discoveries, for example, new drug discoveries from future versions of AlphaFold. In such cases, computational power will be a critical input&#8212;it is not sensible to solely depend upon international markets for a resource which is so central to future prosperity.</p><p>Simply, AI datacentres are an asymmetric bet&#8212;if the AI &#8216;bulls&#8217; are correct, then it is crucial the UK has datacentres for future growth and security, and it is crucial the UK expands its compute industry. If the &#8216;AI bears&#8217; turn out to be correct, the rollout of AI systems will be a multi-decade-long integration of a general-purpose technology and so operating the AI infrastructure will provide jobs and tax revenues for public services; as well as spillovers in knowing how to build cheap power generation in the UK. As private investors are providing the capital, there is no scarce resource being consumed by creating the conditions for them to invest. The bar for deciding to pursue AI datacentres and nuclear power generation is low, as the UK needs growth so dearly.</p><h2>6. Nuclear power is slow and expensive to build, but it could be cheaper and faster</h2><p>UK nuclear power construction is <em>really </em>slow and expensive.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-27" href="#footnote-27" target="_self">27</a> No private investor who wants to power an AI datacentre would choose to build a nuclear power plant in the UK, at present. However, there is a truly remarkable amount of low-hanging fruit to be picked, to become internationally competitive.</p><p>In this section we diagnose the reasons for the high costs, and <a href="https://inferencemagazine.substack.com/getting-ai-datacentres-in-the-uk#7.-The-UK-should-create-'Special-Compute-Zones'">in a later section</a>, we propose a reform package based on this diagnosis.</p><p>To set the scene:</p><ul><li><p>Hinkley Point C is forecast to cost <a href="https://datawrapper.dwcdn.net/U9bFA/1/">&#163;10 million per MW</a>, which is<strong> 4.5 times more expensive</strong> than South Korea, <a href="https://www.samdumitriu.com/p/infrastructure-costs-nuclear-edition?r=1iuptd&amp;utm_campaign=post&amp;utm_medium=web&amp;triedRedirect=true">&#163;2.24 million per MW</a>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-28" href="#footnote-28" target="_self">28</a></p></li><li><p>Construction for Hinkley Point C has been delayed to <strong>14 years</strong>, from nine years planned. Sizewell C is due to take <strong>12 to 15 years</strong> to build. On the contrary, the median time to build a nuclear reactor since 1990 has been under six <a href="https://www.sustainabilitybynumbers.com/p/nuclear-construction-time">years</a>. Between 1970 and 2009, Japan built 60 nuclear power plants in a <a href="https://jackdevanney.substack.com/p/nuclear-power-is-too-slow">median time of 3.8 years</a></p></li><li><p>Hinkley Point C took <strong>six years</strong> to progress from initial consultation to final approval. The consultation process for Sizewell C began <strong>12 years</strong> ago, and EDF will make a final construction decision in Spring 2025. By comparison, France and France took just <strong>three years</strong> and <strong>four and a half years</strong> respectively to approve a plant with the same reactor (the European Pressurised Reactor, or EPR-1600.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-29" href="#footnote-29" target="_self">29</a></p></li><li><p>Three UK projects have been <strong>abandoned</strong> in the pre-construction phase <strong>since 2018</strong> because of financing concerns.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-30" href="#footnote-30" target="_self">30</a></p></li></ul><h4>The total cost of nuclear power is halved if you can borrow cheaply</h4><p>The most important thing to understand about nuclear power projects is that the cost of capital dominates the overall cost of the project. Interest was <strong>approximately two thirds</strong> of the cost of Hinkley Point C.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-31" href="#footnote-31" target="_self">31</a> <a href="https://www.generationatomic.org/the-hinkley-point-c-case-is-nuclear-energy-expensive/">This report</a> estimates the breakdown of Hinkley Point C costs as follows:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-Aua!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7fecd79-7cb6-4cbd-b0ea-953408b7f174_792x589.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-Aua!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7fecd79-7cb6-4cbd-b0ea-953408b7f174_792x589.png 424w, https://substackcdn.com/image/fetch/$s_!-Aua!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7fecd79-7cb6-4cbd-b0ea-953408b7f174_792x589.png 848w, https://substackcdn.com/image/fetch/$s_!-Aua!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7fecd79-7cb6-4cbd-b0ea-953408b7f174_792x589.png 1272w, https://substackcdn.com/image/fetch/$s_!-Aua!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7fecd79-7cb6-4cbd-b0ea-953408b7f174_792x589.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-Aua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7fecd79-7cb6-4cbd-b0ea-953408b7f174_792x589.png" width="792" height="589" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e7fecd79-7cb6-4cbd-b0ea-953408b7f174_792x589.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:589,&quot;width&quot;:792,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!-Aua!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7fecd79-7cb6-4cbd-b0ea-953408b7f174_792x589.png 424w, https://substackcdn.com/image/fetch/$s_!-Aua!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7fecd79-7cb6-4cbd-b0ea-953408b7f174_792x589.png 848w, https://substackcdn.com/image/fetch/$s_!-Aua!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7fecd79-7cb6-4cbd-b0ea-953408b7f174_792x589.png 1272w, https://substackcdn.com/image/fetch/$s_!-Aua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe7fecd79-7cb6-4cbd-b0ea-953408b7f174_792x589.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The cost of the electricity is very sensitive to the cost of borrowing for construction: a 2020 report by the International Energy Agency uses a prototypical EPR-1600 in France, and says when the cost of capital is 3%, instead of 10%, the levelised price of energy is reduced by more than half (53%).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-32" href="#footnote-32" target="_self">32</a></p><p>Why does this matter for our purposes? There are two implications:&nbsp;</p><ol><li><p><strong>Speed matters</strong>; not just because the energy generation can begin, but because it means borrowing can stop.</p></li><li><p><strong>Certainty matters</strong>; when investors perceive a project to be less risky, the more cheaply they provide their capital. <a href="https://illuminem.com/illuminemvoices/nuclear-economics-lessons-from-lazard-to-hinkley-point-c">This report</a> said EDF were forecasting a 9% return on their capital at Hinkley Point C.</p></li></ol><p>Any reforms to planning and regulation not only save money directly through simplification, but they also reduce project risk and timelines, thereby leading to indirect savings on interest payments.</p><h4>Construction costs can be halved again by building reactors &#8216;in fleets&#8217;</h4><p>South Korea is able to build nuclear so cheaply because it builds reactors &#8216;in fleets&#8217;, where it repeats the same reactor design of&nbsp; reduces its cost of nuclear power by building &#8216;fleets&#8217; of 8 to 12 times. This repetition creates <strong>&#8216;learning&#8217; between projects</strong>. (Learning describes the cost declines driven by the <a href="https://ourworldindata.org/learning-curve">cumulative experience of having done something before</a>.) The clearest example of learning is performing technical tasks better. Because nuclear power plant projects are so large, these gains even exist within projects: EDF has claimed that welding for the second reactor at Hinkley Point C is happening <a href="https://www.samdumitriu.com/p/how-to-get-new-nuclear-built-faster">twice as quickly.</a> But an equally important type of learning comes from the developers knowing what the regulators want&#8212;<em>construction </em>is only a small fraction of the activity to start a nuclear reactor, lots of effort is spent on quality assurance and safety. When a reactor is repeated multiple times, the mutual understanding between regulators and developers transfers across projects.</p><p>Fleets also support <strong>supply chain certainty. </strong>The nuclear supply chain requires higher quality assurance standards and more intensive component testing than ordinary industrial projects, which often necessitates a separate supply chain. If there is a large number of reactors to be built, which have been approved previously, the nuclear supply chain could produce components without needing to specify in which specific reactor the parts will be used. The same applies for the <strong>supply chain of skills</strong>. When there is a clear pipeline of construction projects, investing the time to become a nuclear welder is a sensible career choice, as it will raise wages for the long term.&nbsp;</p><p><strong>The UK does the opposite of building in fleets.</strong> The last UK nuclear project to be completed was Sizewell B, in 1995. This reactor was a Pressurised Water Reactor, and it would be 21 years before construction on our next nuclear reactor began at Hinkley Point C. This was a different design, the EPR-1600. As we&#8217;ve noted, this basic reactor design had been used previously in France and Finland, however EDF <a href="https://www.world-nuclear-news.org/Articles/EDF-announces-Hinkley-Point-C-delay-and-big-rise-i">has said</a> the Office for Nuclear Regulation (ONR) required <strong>7,000 design changes</strong> including 25% more concrete and 35% more steel for the reactor to be approved in the UK. The ONR <a href="https://view.officeapps.live.com/op/view.aspx?src=https%3A%2F%2Fwww.onr.org.uk%2Fmedia%2Fqr2ifif4%2Fonrs-regulatory-influence-on-the-epr-design-in-the-uk.docx&amp;wdOrigin=BROWSELINK">disputes this</a>, but irrespective of who was responsible, when the reactor becomes increasingly <em>dissimilar</em>, it is evermore difficult to transfer learning from the previous sites.</p><p>The combination of low levels of regulatory certainty, and no clear precedents for the one-of-a-kind reactor, mean that the supply chain might not have the confidence to justify preparing components until after the final construction decision is made. Likewise, with large gaps between projects, the &#8216;supply chain of skills&#8217; is weaker: workers have lower incentives to develop the skills required for nuclear projects, and move away to other professions. For example, the delays to the approval of Sizewell C mean that nuclear welders will be unable to transition from Hinkley Point C directly into a new project.</p><h4>The nuclear approval process is <a href="https://en.wikipedia.org/wiki/Vetocracy">vetocratic</a>&#8212;there is no positive force in the system to push back against time delays and cost increases.</h4><p>The next largest contributor to slow and expensive nuclear projects is the diffusion of state responsibility for approvals. For a new nuclear power plant to be approved:</p><ul><li><p>The <strong>Office for Nuclear Regulation</strong> must grant a Nuclear Site Licence, which covers the location, the technology, and operation against accidents. The ONR is an independent statutory authority, and a public corporation in the <strong>Department for Work and Pensions</strong>.</p></li><li><p>The <strong>Environment Agency</strong> must grant an Environmental Permit to cover the environmental effects of operation, if the reactor is in England; and in tandem with the <strong>respective devolved administration </strong>if it is elsewhere.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-33" href="#footnote-33" target="_self">33</a></p></li><li><p>The Secretary of State in the <strong>Department for the Environment, Food, and Rural Affairs</strong> must confirm regulatory justification, which states that the benefits of using ionising radiation outweigh the costs.</p></li><li><p>The Secretary of State for the <strong>Department for Energy Security and Net Zero</strong> must approve a Development Consent Order (DCO), which involves multiple rounds of consultation and an Environmental Impact Assessment (EIA).</p></li></ul><p>The ONR is an independent statutory authority, responsible for nuclear fission safety. This means its mandate and responsibility is to prevent accidents that could be <em>caused by nuclear power plants. </em>The ONR has <strong>no authority or responsibility, to weigh the counterfactual risks from not building nuclear power</strong>: for example, the approximately <a href="https://conference.nber.org/conf_papers/f205791.pdf">33 million life-years lost in the UK</a> due to air pollution since nuclear power plant construction was slowed following Chernobyl, or the environmental damage from ongoing greenhouse gas emissions, or the <a href="https://www.economist.com/graphic-detail/2023/05/10/expensive-energy-may-have-killed-more-europeans-than-covid-19-last-winter">impact of high energy prices on people</a> or businesses, or any manner of other challenges.</p><p>The incentive and responsibility of the ONR is to minimise the risk of accidents <em>from nuclear. </em>The global standard for safety regulation, required in all industries in the UK, is that the risk of ionising radiation exposure is &#8216;As Low As Reasonably Practicable&#8217; (nb. in some contexts this might be &#8216;As Low As Reasonably Achievable&#8217;). Because the ONR is not set up to balance aims, such reasonableness is defined as anything which can improve reactor safety, until a measure can be proven to be &#8216;grossly disproportionate&#8217;.</p><p>Similarly, the consequences of long and uncertain Environmental Impact Assessments are not considered. The EIA for Sizewell C <a href="https://www.gov.uk/government/publications/getting-great-britain-building-again-speeding-up-infrastructure-delivery/getting-great-britain-building-again-speeding-up-infrastructure-delivery">was 44,260 pages</a>; and for Hinkley Point C it was 31,401 pages. The issues of EIAs are not unique to nuclear power, and so we leave these to other sources, however, <a href="https://www.samdumitriu.com/p/visiting-the-worlds-most-expensive">a report</a> by Sam Dumitriu claims that EDF have spent, &#8220;hundreds of millions [of pounds]&#8221;, at Hinkley Point C to install underwater speakers, in order to deter roughly 112 fish from entering the water cooling system.&nbsp;</p><p>Unlike the UK, in South Korea, the Nuclear Safety and Security Commission reports directly to the Prime Minister.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-34" href="#footnote-34" target="_self">34</a> <a href="https://jackdevanney.substack.com/p/alaras-not-so-bad-what-about-korea">This report</a> suggests that political oversight changes the incentive equilibrium for regulators, to more appropriately balance the costs and benefits of incremental safety regulations. Though the ONR is independent, it is also a public corporation within the Department for Work and Pensions, and therefore there is less Ministerial interest or bandwidth for seeing that nuclear power gets built.</p><h4>&#8216;Regulatory justification&#8217; has been misapplied</h4><p>&#8216;Regulatory justification&#8217; is a requirement that stems from a 1996 EU directive that stipulates the benefits of ionising radiation for the production of energy must outweigh the costs. This does not seem, in principle, to be a bad idea&#8212;<em>who </em>would be <em>for </em>using ionising radiation where the costs exceeded the benefits? However, the requirement applies to each &#8216;practice&#8217;, which is an instance of the use of ionising radiation. It is currently interpreted that <em>each reactor design </em>is its own practice which must be separately assessed for regulatory justification. There are good legal arguments that nuclear power, or broad characteristics such as using low enriched uranium and light water as coolant and moderator, should be a single practice for which &#8216;regulatory justification&#8217; is established once and for all.</p><p>Because of &#8216;functional separation&#8217;, authority for this decision sits with the Department for the Environment, Food, and Rural Affairs, and the decision takes two years. This is duplicative, because the purpose of the planning process is to weigh the relative merits of a new project. France, Finland, and Sweden incorporate the &#8216;regulatory justification&#8217; into their planning decisions.</p><p>This superfluous step increases project uncertainty and duration and therefore raises the total costs of the project by causing longer and more risky borrowing.</p><p>To summarise, there are many actors in the system who can say &#8216;no&#8217;, or who might add incremental delays and cost increases to nuclear power plant construction, which amounts to a <em>de facto </em>&#8216;no&#8217;; but there is no positive force which pushes back against slowness, expensiveness, and the counterfactual damages of the two. With an approach that weighs cost and benefits, including the knock-on impacts to speed and certainty, the UK can build an internationally competitive nuclear planning and regulatory regime.</p><h2>7. The UK should create &#8216;Special Compute Zones&#8217;</h2><p>The purpose of this reform proposal is to solve an incongruence:</p><ul><li><p>The UK&#8217;s AI datacentre capacity is an imperative for economic security, growth in former places of industry, and long-term prosperity; as this previous section set out.</p></li><li><p>But the planning and regulatory approval process for building new AI datacentres and associated power cannot permit the amount of construction, at the speed required.</p></li></ul><p>Below is our proposal on how this might be resolved.</p><p>Within &#8216;Special Compute Zones&#8217;, there is an alternative planning and regulatory approval process for nuclear reactors, AI datacentres, and the transmission and networking infrastructure they require. The goal is to provide the <strong>certainty, speed, and hence, the opportunity to be cheap,</strong> that would make the UK the most competitive place in the world to build AI datacentres.</p><p>Within the Zones, there would be &#8216;deemed consent&#8217;&#8212;meaning the default answer to construction is &#8216;yes&#8217;&#8212;and permission to construct would have to be challenged within three months of an application. Ordinarily, planning decisions weigh the relative merits of each project; but within a &#8216;Special Compute Zone&#8217;, it would be decided that <em>by creating a zone, </em>this cost-benefit has already been considered, and therefore approval would depend on a &#8216;condition-based approach&#8217;. This means that if a developer can prove the project meets particular standards, it goes ahead. There is precedent for location-based policies from Spain and the EU.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-35" href="#footnote-35" target="_self">35</a> In Spain, the Government passed a decree which allowed renewable projects to forego the Environmental Impact Assessments, so long as the project met some <a href="https://www.energy-transitions.org/wp-content/uploads/2023/01/Barriers_PP_GovernmentST_vFinal.pdf">conditions</a>:</p><ul><li><p>Wind and solar projects are below 75 MW and 150 MW respectively.</p></li><li><p>Projects are in areas of low or moderate environmental sensitivity.</p></li><li><p>Grid connection lines are not longer than 15km or above 220kV.</p></li><li><p>Authorities do not lodge an objection within two months.</p></li></ul><p><a href="https://www.energy-transitions.org/wp-content/uploads/2023/01/Barriers_PP_GovernmentST_vFinal.pdf">This report</a> notes the change doubled the speed of projects, and increased the forecast of solar construction by 13GW for 2030.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-36" href="#footnote-36" target="_self">36</a><br><br>The EU has mandated &#8220;Renewable Acceleration Areas&#8221;, as of September 2023, which require Member States to designate at least one area by February 2026. To implement them, Member States <a href="https://energy.ec.europa.eu/document/download/af3927a5-3b82-42f0-8954-7b9fdc567e43_en?filename=SWD_2024_333_2_EN_autre_document_travail_service_part1_v1.pdf">prepare</a>, &#8220;a mitigation &#8216;rulebook&#8217; consisting of a set of rules on mitigation measures to adopt in the specific area, aimed at avoiding or where not possible significantly reducing the environmental impacts resulting from the installation of projects in those areas.&#8221;</p><p>The Secretary of State in the Department for Science, Information, and Technology would be able to <strong>provide a single sign-off</strong> for the projects. This means nuclear power, AI datacentres, and networking and transmission infrastructure would be approved <em>together</em>, to improve project certainty. As with renewable zones, developers would apply to the Secretary of State, who would have three months to object to the application for construction, after which the project would automatically have permission. This aligns the incentives of governments to quickly respond to applications, as the average consideration period of a Development Consent Order is 22 months.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-37" href="#footnote-37" target="_self">37</a> Furthermore, a non-objection procedure, rather than a positive decision on each site, reduces the risk of judicial review. Judicial review proceedings could slow construction by roughly two years; and so developers are likely to want clarity that everything possible has been done to avoid judicial review. In order to further reduce the risk of judicial review, the primary legislation for Special Compute Zones could:</p><ul><li><p>Specify the grounds on which a challenge to the Act can be brought.</p></li><li><p>Exclude &#8220;oral renewal&#8221; for judicial review.</p></li><li><p>Ensure reviews are brought for <em>procedural </em>reasons and not ones based on the principle, policy or merits of the proposal.</p></li></ul><p>One area for further investigation is to what extent all of the application process needs to be frontloaded before construction can begin, or whether this can be parallelised.&nbsp; For example, some environmental permitting and mitigation (i.e. dealing with environmental effects of plant operation) might be performed concurrently with construction, subject to a condition that the plant cannot start operation until they have been addressed. There is international precedent for this: nuclear power developers in the US can opt to licence their reactors through &#8216;Part 50&#8217; of the U.S. Code of Federal Regulations, which grants them separate construction and operating licences. (Ordinarily developers use Part 52, which grants construction and operation licences together.) The developer takes on some risk that they aren&#8217;t granted an operating licence, when they begin construction without.&nbsp;</p><p>The &#8220;Special Compute Zones&#8221; will need to permit nuclear power operators to make &#8216;behind the meter&#8217; power purchase agreements with datacentre operators. The grid network fees are 20-25% of the cost of grid power, and so it is necessary to make the UK a competitive place to build.</p><h4>How should the Zones be designated?</h4><p>The primary legislation for Special Compute Zones could designate <em>all </em>former nuclear, coal, and natural gas power plant sites, former steel plants, and ports as Special Compute Zones, and give the Secretary of State the power to de-designate or designate individual sites or classes of sites. The primary legislation would need to designate sufficiently broad <em>classes of sites </em>to avoid being a hybrid bill. This is because hybrid bills have a much longer parliamentary process, and therefore would take much longer to pass, and so it would make it much more difficult for the UK to participate in building AI datacentres. By making all of these classes Special Compute Zones through primary legislation, and providing the Secretary of State power to object and de-designate, the risk of judicial review is considerably lower than if the Secretary of State had to make a decision to designate each particular site. This in turn, means it is more likely that AI datacentres would be built.</p><p>The Secretary of State should also be given the power to determine which environmental conditions apply to each zone, based on a high level environmental assessment, so that environmental protection is provided. Compliance with those conditions would, under the Act, be a basis for removing the need for EIA.</p><p>Former sites of energy generation and steel production are especially suitable to be Special Compute Zones as they have grid connections and previously had environmental impacts that are likely equal to or, indeed, much greater than nuclear power plant operation. This improves the ease and strength of argument for streamlining the regulatory and planning process. Ports are suitable for conventional nuclear because bringing materials to the site by boat is substantially easier than to other sites. For example, at Hinkley Point C, wharfs needed to be constructed for delivery to reduce the number of lorries.</p><p>There may be additional options for planning and regulatory approval, worth including as alternative pathways, although they are unlikely to deliver results within the necessary timeframes as the proposal suggested above. For example, the local planning authority could grant consent for new nuclear power and datacentres, provided the application met the conditions required by the Special Compute Zones. To incentivise local governments to allow more power and datacentres, they might be permitted to retain <a href="https://www.sambowman.co/p/one-weird-trick-to-get-data-centres">100% of the business rates in perpetuity</a><strong> </strong>for projects approved by this mechanism.</p><h4>What should the conditions be?</h4><p>For deemed consent to apply, the reactor must have been approved by a recognised international regulator (i.e. the US or the EU). Introducing international recognition would speed up both approval and construction.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-38" href="#footnote-38" target="_self">38</a> Practically, international recognition would allow international teams who have already built fleets of reactors to come to the UK, upskill local workers, and replicate their previous work, saving costs with the lessons they learned during that work.</p><p>The conditions for radiation dose exposure for workers and the public could be taken from the <a href="https://www.legislation.gov.uk/uksi/2017/1075/schedule/3#:~:text=Employees%20and%20trainees%20of%2018%20years%20of%20age%20or%20above&amp;text=(b)the%20limit%20on%20equivalent,mSv%20in%20a%20calendar%20year">Ionising Radiations Regulations 2017</a>.</p><p>Some conditions would relate to environmental impacts of construction and operation, to streamline and replace environmental permitting and impact assessments, and could include options such as requiring the developer to pay into a rewilding fund. This will ensure environmental effects are addressed without creating uncertainty and expensive delays.</p><p>Some conditions would address questions generally addressed in Nuclear Site Licences. For example, there should be requirements about emergency planning in scenarios with technical failure, and population density around the site.</p><p>Crucially, these conditions should differ between conventional nuclear and SMRs. For example, safety requirements should be proportionate to the size of the plant.&nbsp;</p><h4>How should we deal with &#8216;regulatory justification&#8217;?</h4><p>Because regulatory justification is an EU directive, it would either need to be incorporated into a planning decision, per the French, Finnish, and Swedish approach, or it could be addressed by a Regulation 9 decision declaring all classes of nuclear power as a new justified practice or issuing a Regulation 12 determination that nuclear energy is an existing practice due to the operation of Magnox, Advanced Gas Reactors, and Sizewell B before the introduction of the directive.</p><div><hr></div><h3>Technical Appendix</h3><p>The model solves for the optimal solar, wind and gas combination to minimise the levelised wholesale cost of electricity.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-39" href="#footnote-39" target="_self">39</a> The model generated 100 possible sets of 365 days outputs of solar and wind, and optimised the selection over 4000 possible combinations of solar, wind and batteries output.&nbsp;</p><p><strong>Solar</strong></p><p>Solar energy has two sources of variance in the model - one caused by variation in average output across months and the other caused by daily variation caused by cloud cover and similar. Average output varies due to the northern hemisphere tilting away from the sun in winter - which both directly reduces incidence by trigonometry but also indirectly does so by forcing sunlight to pass through thicker air before reaching the Earth&#8217;s surface. Average solar output in December is thus <a href="https://www.otovo.co.uk/blog/photovoltaic-systems/how-much-energy-can-solar-panels-produce/">72%</a> lower than its level in summer - with this effect being much greater than temperature variations imply as air can move around the earth. Cloud cover also reduces output, giving daily variation following a <a href="https://www.redalyc.org/journal/430/43067845004/html/">gamma distribution</a> with shape 3.5. In the UK, solar energy has a capacity factor of <a href="https://en.wikipedia.org/wiki/Solar_power_in_the_United_Kingdom#Solar_PV_installed_capacity_and_generation">10%</a>, a levelised cost of <a href="https://assets.publishing.service.gov.uk/media/6556027d046ed400148b99fe/electricity-generation-costs-2023.pdf">&#163;49/MWh</a>, and 1GW <a href="https://consultations.rochdale.gov.uk/research/solar-farm/supporting_documents/STA%20solar%20farm%20factsheet%20NEW.pdf">occupies</a> 20km<sup>2</sup>. In Texas, the levelised cost ranges from <a href="https://www.pv-magazine.com/2023/04/14/average-solar-lcoe-increases-for-first-time-this-year/">$24 to $96</a>, giving a central estimate of $46/MWh. Solar output in Texas is much less volatile, declining by only <a href="https://shrinkthatfootprint.com/average-solar-production-in-texas-usa/">40%</a> in winter.</p><p><strong>Wind&nbsp;</strong></p><p>Wind energy varies in output due to variations in wind speed, which occur both across days and seasons. During winter, UK wind speeds are <a href="https://www.statista.com/statistics/322789/quarterly-wind-speed-average-in-the-united-kingdom-uk/#:~:text=Quarterly%20average%20wind%20speed%20in%20the%20United%20Kingdom%202010%2D2024&amp;text=Wind%20speed%20averages%20in%20the,an%20average%20of%209.4%20knots.">25% higher</a> than summer, giving it the beneficial property of being weakly negatively correlated to solar. However, the underlying energy stored within wind goes as the cube of wind speed, as each individual molecule&#8217;s kinetic energy goes as the square of wind speed and linearly more air molecules hit the turbine&#8217;s blades per second as wind speed increases. Although wind power in practice hits diminishing returns and eventually decreases in power output in wind speed, for most of the wind speed distribution in the UK we remain in the cubic - and so highly volatile - portion. This is compounded by the high volatility in the underlying distribution - <a href="https://electricajournal.org/Content/files/sayilar/78/1907-1912.pdf">Rayleigh</a>, with a standard deviation of half its mean. In the UK, wind has a capacity factor of <a href="https://energynumbers.info/uk-offshore-wind-capacity-factors">27%</a>, a levelised price of <a href="https://assets.publishing.service.gov.uk/media/6556027d046ed400148b99fe/electricity-generation-costs-2023.pdf">&#163;53/MWh</a> and 1GW of wind <a href="https://www.ft.com/content/d3b8947a-bdb1-445e-80f7-a19b51dd977d">occupies</a> 110km<sup>2</sup>. The cost ranges from <a href="https://en.wikipedia.org/wiki/Wind_power_in_the_United_States">$24 to $75</a>, implying a central estimate of &#163;38/MWh.</p><p><strong>Batteries</strong></p><p>Batteries offer the ability to smooth variable-output processes and provide the constant stream of power needed for efficient datacentre operation. Additionally, as they are primarily used in summer the scale-up of solar over the day still only means 12GWh are required for continuous operation. However, they still remain impractically expensive to participate in balancing on timescales longer than a day, and also remain limited by generally only being able to charge at the same rate as they discharge - meaning that very large amounts of excess generation by renewables across seasons is difficult to exploit. Existing grid-scale storage has a price of <a href="https://www.nrel.gov/docs/fy23osti/85332.pdf">$400/MWh</a> and lasts for at least <a href="https://www.fluxpower.com/blog/lithium-ion-vs.-lead-acid-battery-life#:~:text=The%20minimum%20lifespan%20most%20manufacturers,as%20long%20as%203%2C000%20cycles.">2000 cycles</a>.&nbsp;</p><p><strong>Gas</strong></p><p>Gas has a levelised price of <a href="https://assets.publishing.service.gov.uk/media/6556027d046ed400148b99fe/electricity-generation-costs-2023.pdf">&#163;136/MWh</a>, just over half of which is carbon pricing. The capacity factor is close to 100%, and the area occupied is very small in comparison to renewables. In Texas, the cost is <a href="https://en.wikipedia.org/wiki/Cost_of_electricity_by_source">$39-$68</a>, so a central estimate of &#163;41.&nbsp;</p><p><strong>Results</strong></p><p>In the optimal allocation, 8GW of solar and 0.4GW of wind was procured, at a cost of &#163;106/MWh with 28% of electricity being gas generated. If it was possible to sell to the grid at &#163;30/MWh during periods of renewables surplus (plausibly too high for the unsubsidised price as this is so correlated nationally), then the optimal solar quantity rises to 12GW, the optimal wind quantity stays unchanged on 0.4GW and the carbon intensity drops to 0.17GW, and the levelised cost drops to &#163;98/MWh. The levelised decline is so small in response to the grid connection because the previous allocation only wasted 18% of its total output as curtailed renewables, so without a large adjustment on the renewables side (which are unprofitable without a guaranteed price on sale induced by subsidies) this leads to little change in prices. Given that the grid connection would of course have to be built and maintained, this renders its provision unlikely to reduce overall costs. Total expected deaths/GW-year are 7.06. The cheapest equivalent-emissions option for Texas costs &#163;74/MWh, being composed of 0.7GW of wind and 4.7GW of solar. If Texas faced the same carbon pricing as British gas, the cost remains substantially lower at &#163;89/MWh.&nbsp;</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>This is different from &#8216;ordinary&#8217; computing, which can only do exactly and specifically what its program (a pre-set instruction for information processing) tells it to do. On the contrary, AI systems, more like brains, are able to learn and have flexible and adaptive information processing.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>From the <a href="https://openai.com/index/learning-to-reason-with-llms/">95.6% raw percentage score</a> would typically fall in the 99th percentile of LSAT performance.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>The average energy consumption of a UK resident annually is 4,266kWh, an H100 chip would use 6,132 kWh. The next generation B200 chip would use 10,512kWh over the course of a year (assuming continuous running).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>OfGem has a medium-sized household using 2,700 kWh in electricity annually, then 27 MW of installed power capacity is 87,600 households.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>The UK produced 310 TWh of electricity in 2021, and 100 GW of generation running for 9.12 months would produce 665.18 TWh of power.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>Over the course of a year, a 1GW datacentre will use 8760 gigawatt-hours (GWh) of power, but across both industry and domestic demand, Liverpool used just 1696 GWh in 2022.&nbsp;</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>Sources for <a href="https://simple.wikipedia.org/wiki/Cardiff">Cardiff</a> and <a href="https://www.britannica.com/place/Reading-England">Reading</a> in square km.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>The CfD was &#163;92.50/MWh in 2012 prices, and there has been <a href="https://www.rateinflation.com/inflation-rate/uk-historical-inflation-rate/">54.5% inflation</a> since, meaning 2024 prices are &#163;142.83/MWh.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p>Potentially including the scientific research which goes into the creation of better AI systems, though exploring the details of this possibility is beyond the scope of this report.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-10" href="#footnote-anchor-10" class="footnote-number" contenteditable="false" target="_self">10</a><div class="footnote-content"><p>The definition of superintelligence is necessarily weaker, as we are less confident about what such a system would look like.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-11" href="#footnote-anchor-11" class="footnote-number" contenteditable="false" target="_self">11</a><div class="footnote-content"><p><a href="https://ourworldindata.org/grapher/test-scores-ai-capabilities-relative-human-performance">https://ourworldindata.org/grapher/test-scores-ai-capabilities-relative-human-performance</a>&nbsp;</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-12" href="#footnote-anchor-12" class="footnote-number" contenteditable="false" target="_self">12</a><div class="footnote-content"><p><a href="https://arxiv.org/abs/2309.11690">This paper</a> reviews the arguments for and against, and <a href="https://situational-awareness.ai/from-agi-to-superintelligence/">this report</a> by a former OpenAI employee makes the case for how AI systems would soon become capable of supporting explosive growth, while noting potential bottlenecks.&nbsp;</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-13" href="#footnote-anchor-13" class="footnote-number" contenteditable="false" target="_self">13</a><div class="footnote-content"><p>Technically, language models are trained to predict &#8216;sub-word units&#8217; called &#8216;tokens&#8217;, hence why this task is sometimes referred to as &#8216;next token prediction&#8217;.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-14" href="#footnote-anchor-14" class="footnote-number" contenteditable="false" target="_self">14</a><div class="footnote-content"><p>It is a common discussion point to debate whether a neural network&#8217;s prediction constitutes &#8216;true&#8217; understanding. Given the enormous downstream capabilities of the models <a href="https://docs.google.com/document/d/1nEyVnyx3DWb1N9Lg4H5f-xfByfpMB-kmOquxo1AX5mY/edit?tab=t.0#heading=h.w9g0142z7nyz">already mentioned</a>, it seems to us to be quite clear important things are happening inside the models, and semantic debates have tended to take the oxygen from more&nbsp;pressing questions about how to integrate powerful systems into society.&nbsp;</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-15" href="#footnote-anchor-15" class="footnote-number" contenteditable="false" target="_self">15</a><div class="footnote-content"><p>It is worth noting that improving the energy and computational efficiency of hardware will induce further demand for hardware. It is a <a href="https://en.wikipedia.org/wiki/Jevons_paradox">Jevons Paradox</a>. (Efficiency might usually mean <em>reduction</em>, but this is not the case.)</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-16" href="#footnote-anchor-16" class="footnote-number" contenteditable="false" target="_self">16</a><div class="footnote-content"><p> <a href="https://epochai.org/blog/trends-in-machine-learning-hardware">https://epochai.org/blog/trends-in-machine-learning-hardware</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-17" href="#footnote-anchor-17" class="footnote-number" contenteditable="false" target="_self">17</a><div class="footnote-content"><p><a href="https://epochai.org/blog/training-compute-of-frontier-ai-models-grows-by-4-5x-per-year#language-models-caught-up-to-the-frontier-around-2020">https://epochai.org/blog/training-compute-of-frontier-ai-models-grows-by-4-5x-per-year#language-models-caught-up-to-the-frontier-around-2020</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-18" href="#footnote-anchor-18" class="footnote-number" contenteditable="false" target="_self">18</a><div class="footnote-content"><p>The average energy consumption of a UK resident annually is 4,266kWh, an H100 chip would use 6,132 kWh; and a B200 chip would use 10,512kWh over the course of a year (assuming continuous running).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-19" href="#footnote-anchor-19" class="footnote-number" contenteditable="false" target="_self">19</a><div class="footnote-content"><p><a href="https://epochai.org/blog/can-ai-scaling-continue-through-2030#the-current-trend-of-ai-power-demand">https://epochai.org/blog/can-ai-scaling-continue-through-2030#the-current-trend-of-ai-power-demand</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-20" href="#footnote-anchor-20" class="footnote-number" contenteditable="false" target="_self">20</a><div class="footnote-content"><p>There were 85492 households in York, at <a href="https://www.ons.gov.uk/datasets/TS041/editions/2021/versions/2">the most recent census</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-21" href="#footnote-anchor-21" class="footnote-number" contenteditable="false" target="_self">21</a><div class="footnote-content"><p>Context: the UK produced 310 TWh of electricity in 2021, and 100 GW of generation running for 9.12 months would produce 665.18 TWh of power.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-22" href="#footnote-anchor-22" class="footnote-number" contenteditable="false" target="_self">22</a><div class="footnote-content"><p>The growth rate in &#8220;AI datacentre critical IT power&#8221; between 2026 and 2027, and 2027 and 2028, <a href="https://www.semianalysis.com/p/ai-datacenter-energy-dilemma-race">using this SemiAnalysis report</a>, is 46.9% and 36.2% respectively.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-23" href="#footnote-anchor-23" class="footnote-number" contenteditable="false" target="_self">23</a><div class="footnote-content"><p>There is some variability based on operational practices&#8212;per <a href="https://www.spectator.co.uk/article/could-the-koreans-save-angleseys-nuclear-power-project/#:~:text=Can%20the%20South%20Koreans%20help,another%20reason%20to%20move%20fast.">this article</a>, &#8220;South Korean plants are five times less likely to lose capacity due to unplanned outages than UK plants&#8221;.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-24" href="#footnote-anchor-24" class="footnote-number" contenteditable="false" target="_self">24</a><div class="footnote-content"><p>&nbsp;<a href="https://www.visualcapitalist.com/how-does-u-s-electricity-generation-change-over-one-week/">https://www.visualcapitalist.com/how-does-u-s-electricity-generation-change-over-one-week/</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-25" href="#footnote-anchor-25" class="footnote-number" contenteditable="false" target="_self">25</a><div class="footnote-content"><p><a href="https://ourworldindata.org/safest-sources-of-energy">https://ourworldindata.org/safest-sources-of-energy</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-26" href="#footnote-anchor-26" class="footnote-number" contenteditable="false" target="_self">26</a><div class="footnote-content"><p>Why Nuclear Power Has Been A Flop, https://gordianknotbook.com/, p.171; via this <a href="https://worksinprogress.co/issue/taming-the-stars/">source</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-27" href="#footnote-anchor-27" class="footnote-number" contenteditable="false" target="_self">27</a><div class="footnote-content"><p>This section draws on the work by Britain Remade, a think tank for economic growth. We are grateful for their high-quality work on nuclear construction that has informed a lot of this proposal, we&nbsp; link to their work where appropriate.&nbsp;</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-28" href="#footnote-anchor-28" class="footnote-number" contenteditable="false" target="_self">28</a><div class="footnote-content"><p>Note that EDF is paying for construction, not the taxpayer. The government agreed a CfD with EDF to provide energy at &#163;92.50/MWh in 2012 prices.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-29" href="#footnote-anchor-29" class="footnote-number" contenteditable="false" target="_self">29</a><div class="footnote-content"><div class="embedded-post-wrap" data-attrs="{&quot;id&quot;:139426431,&quot;url&quot;:&quot;https://www.samdumitriu.com/p/infrastructure-costs-nuclear-edition&quot;,&quot;publication_id&quot;:219115,&quot;publication_name&quot;:&quot;Notes on Growth&quot;,&quot;publication_logo_url&quot;:null,&quot;title&quot;:&quot;Infrastructure Costs: Nuclear Edition&quot;,&quot;truncated_body_text&quot;:&quot;Britain used to lead the world in nuclear power. This is the country that split the atom, built the world&#8217;s first full-scale nuclear power station, and then proceeded to build nine more in the decade that followed. When Calder Hall was opened, Lord Privy Seal, Richard Butler, noted &#8220;It may be that after 1965 every new power station being built will be a&#8230;&quot;,&quot;date&quot;:&quot;2023-12-04T13:38:19.100Z&quot;,&quot;like_count&quot;:24,&quot;comment_count&quot;:8,&quot;bylines&quot;:[{&quot;id&quot;:92132401,&quot;name&quot;:&quot;Sam Dumitriu&quot;,&quot;handle&quot;:&quot;samdumitriu&quot;,&quot;previous_name&quot;:null,&quot;photo_url&quot;:&quot;https://bucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com/public/images/a89ca9d2-0dad-4ce8-9412-21bc724042e3_2204x2939.jpeg&quot;,&quot;bio&quot;:&quot;Head of Policy at Britain Remade.&quot;,&quot;profile_set_up_at&quot;:&quot;2022-09-30T21:18:41.126Z&quot;,&quot;publicationUsers&quot;:[{&quot;id&quot;:225610,&quot;user_id&quot;:92132401,&quot;publication_id&quot;:219115,&quot;role&quot;:&quot;admin&quot;,&quot;public&quot;:true,&quot;is_primary&quot;:false,&quot;publication&quot;:{&quot;id&quot;:219115,&quot;name&quot;:&quot;Notes on Growth&quot;,&quot;subdomain&quot;:&quot;samdumitriu&quot;,&quot;custom_domain&quot;:&quot;www.samdumitriu.com&quot;,&quot;custom_domain_optional&quot;:false,&quot;hero_text&quot;:&quot;Markets, tech and policy.&quot;,&quot;logo_url&quot;:null,&quot;author_id&quot;:92132401,&quot;theme_var_background_pop&quot;:&quot;#009B50&quot;,&quot;created_at&quot;:&quot;2020-11-20T17:47:01.659Z&quot;,&quot;rss_website_url&quot;:null,&quot;email_from_name&quot;:null,&quot;copyright&quot;:&quot;Sam Dumitriu&quot;,&quot;founding_plan_name&quot;:null,&quot;community_enabled&quot;:true,&quot;invite_only&quot;:false,&quot;payments_state&quot;:&quot;disabled&quot;,&quot;language&quot;:null,&quot;explicit&quot;:false,&quot;is_personal_mode&quot;:false}}],&quot;twitter_screen_name&quot;:&quot;Sam_Dumitriu&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null},{&quot;id&quot;:100030867,&quot;name&quot;:&quot;Ben Hopkinson&quot;,&quot;handle&quot;:&quot;benhopkinson&quot;,&quot;previous_name&quot;:null,&quot;photo_url&quot;:&quot;https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe37983bc-7c22-41b9-b471-070db085ee0e_886x886.jpeg&quot;,&quot;bio&quot;:&quot;Head of Research at Britain Remade&quot;,&quot;profile_set_up_at&quot;:&quot;2023-08-24T09:30:34.573Z&quot;,&quot;is_guest&quot;:true,&quot;bestseller_tier&quot;:null,&quot;primaryPublicationId&quot;:1899416,&quot;primaryPublicationName&quot;:&quot;Yes and Grow&quot;,&quot;primaryPublicationUrl&quot;:&quot;https://benhopkinson.substack.com&quot;,&quot;primaryPublicationSubscribeUrl&quot;:&quot;https://benhopkinson.substack.com/subscribe?&quot;}],&quot;utm_campaign&quot;:null,&quot;belowTheFold&quot;:true,&quot;type&quot;:&quot;newsletter&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="EmbeddedPostToDOM"><a class="embedded-post" native="true" href="https://www.samdumitriu.com/p/infrastructure-costs-nuclear-edition?utm_source=substack&amp;utm_campaign=post_embed&amp;utm_medium=web"><div class="embedded-post-header"><span></span><span class="embedded-post-publication-name">Notes on Growth</span></div><div class="embedded-post-title-wrapper"><div class="embedded-post-title">Infrastructure Costs: Nuclear Edition</div></div><div class="embedded-post-body">Britain used to lead the world in nuclear power. This is the country that split the atom, built the world&#8217;s first full-scale nuclear power station, and then proceeded to build nine more in the decade that followed. When Calder Hall was opened, Lord Privy Seal, Richard Butler, noted &#8220;It may be that after 1965 every new power station being built will be a&#8230;</div><div class="embedded-post-cta-wrapper"><span class="embedded-post-cta">Read more</span></div><div class="embedded-post-meta">2 years ago &#183; 24 likes &#183; 8 comments &#183; Sam Dumitriu and Ben Hopkinson</div></a></div></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-30" href="#footnote-anchor-30" class="footnote-number" contenteditable="false" target="_self">30</a><div class="footnote-content"><p> Wylfa, Moorside, and Oldbury.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-31" href="#footnote-anchor-31" class="footnote-number" contenteditable="false" target="_self">31</a><div class="footnote-content"><p>There is not a publicly available breakdown from EDF, but the International Energy Agency published a report in 2020 which said the overnight capital cost (i.e. exl. interest) of building a European Pressurised Reactor (the reactor used at Hinkley Point C) is $4013/kWe in 2018 USD. Exchanging to 2018 GBP (<a href="https://www.oecd.org/en/data/indicators/exchange-rates.html?oecdcontrol-00b22b2429-var3=2018&amp;oecdcontrol-38c744bfa4-var1=GBR%7CUSA">at 1:0.75</a>) and adjusting for 2024 prices (<a href="https://www.rateinflation.com/inflation-rate/uk-historical-inflation-rate/">26% inflation</a>), implies an overnight capital cost of &#163;6.26bn GBP 2024 for each reactor. The total <a href="https://www.world-nuclear-news.org/Articles/EDF-announces-Hinkley-Point-C-delay-and-big-rise-i">project cost</a>, <a href="https://www.rateinflation.com/inflation-rate/uk-historical-inflation-rate/">adjusting for inflation</a> is &#163;41.31 billion to &#163;45.31 billion in GBP 2024. Which suggests that 69.7% to 72.3% of capital costs is interest. From conversations with experts, this is on track with other estimates, though perhaps on the higher end.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-32" href="#footnote-anchor-32" class="footnote-number" contenteditable="false" target="_self">32</a><div class="footnote-content"><p><a href="https://iea.blob.core.windows.net/assets/ae17da3d-e8a5-4163-a3ec-2e6fb0b5677d/Projected-Costs-of-Generating-Electricity-2020.pdf">https://iea.blob.core.windows.net/assets/ae17da3d-e8a5-4163-a3ec-2e6fb0b5677d/Projected-Costs-of-Generating-Electricity-2020.pdf</a>, p.59</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-33" href="#footnote-anchor-33" class="footnote-number" contenteditable="false" target="_self">33</a><div class="footnote-content"><p>Natural Resources Wales, the Scottish Environment Protection Agency, or Northern Ireland Environment Agency.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-34" href="#footnote-anchor-34" class="footnote-number" contenteditable="false" target="_self">34</a><div class="footnote-content"><p>The organisation of the state in South Korea is different to the UK; the President takes an active role in administration.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-35" href="#footnote-anchor-35" class="footnote-number" contenteditable="false" target="_self">35</a><div class="footnote-content"><p>h/t to Sam Dumitriu for <a href="https://www.samdumitriu.com/p/how-spain-eliminated-environmental">highlighting this</a> on his Substack.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-36" href="#footnote-anchor-36" class="footnote-number" contenteditable="false" target="_self">36</a><div class="footnote-content"><p>Bloomberg NEF&#8217;s 2030 solar forecast was raised from 73GW in April 2022 to 86GW in October 2022.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-37" href="#footnote-anchor-37" class="footnote-number" contenteditable="false" target="_self">37</a><div class="footnote-content"><p>Again, h/t to Sam Dumitriu for <a href="https://www.samdumitriu.com/p/why-britain-struggles-to-build-infrastructure">highlighting this</a> on his Substack.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-38" href="#footnote-anchor-38" class="footnote-number" contenteditable="false" target="_self">38</a><div class="footnote-content"><p>Once again, h/t to Sam Dumitriu for <a href="https://www.samdumitriu.com/p/how-to-get-new-nuclear-built-faster">highlighting this</a> on his Substack</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-39" href="#footnote-anchor-39" class="footnote-number" contenteditable="false" target="_self">39</a><div class="footnote-content"><p>Implicitly this assumes that exclusively cogeneration occurs as this allows transportation costs to be ignored - meaning that wind is considered exclusively onshore not offshore.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[Coming soon]]></title><description><![CDATA[This is Inference.]]></description><link>https://inferencemagazine.substack.com/p/coming-soon</link><guid isPermaLink="false">https://inferencemagazine.substack.com/p/coming-soon</guid><dc:creator><![CDATA[Inference]]></dc:creator><pubDate>Fri, 15 Nov 2024 00:39:56 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!9o2z!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c61df20-b545-4c7b-9acb-75940496383f_1010x1010.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>This is Inference.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://inferencemagazine.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://inferencemagazine.substack.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item></channel></rss>